<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Manual and Automated Labeling of Web User Interfaces for User Behavior Models</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Anna</forename><surname>Stepanova</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Novosibirsk State Technical University</orgName>
								<address>
									<postCode>630073</postCode>
									<settlement>Novosibirsk</settlement>
									<country key="RU">Russia</country>
								</address>
							</affiliation>
						</author>
						<author role="corresp">
							<persName><forename type="first">Maxim</forename><surname>Bakaev</surname></persName>
							<email>bakaev@corp.nstu.ru</email>
							<affiliation key="aff0">
								<orgName type="institution">Novosibirsk State Technical University</orgName>
								<address>
									<postCode>630073</postCode>
									<settlement>Novosibirsk</settlement>
									<country key="RU">Russia</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Manual and Automated Labeling of Web User Interfaces for User Behavior Models</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">0112380111AF23D6C81AAE20FB4E4F17</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T05:32+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Image labeling</term>
					<term>human-computer interfaces</term>
					<term>machine learning</term>
					<term>linear regression</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The article contrasts manual and automated identification of elements in images of web user interfaces (UIs), which is essential for machine learning (ML) models that describe user behavior. We consider the principal advantages and disadvantages of the two methods and compare linear regression models. The constructed ML models describe users' subjective perception of web UIs in such dimensions as complexity, aesthetics and ordering. Somehow unexpectedly, the resulting R 2 s of models built with certain factors obtained from automated labeling turned out to be slightly higher. Particularly, shares of text and images in the web UI, as well as the sizes of the elements, were rather influential. We believe that the main disadvantage of the manual labeling is the human factor, as mistakes made by the labelers and diversity of their outcome affect the quality of the models. In turn, the automated process has a number of drawbacks that must be taken into account and that we discuss in the paper. The results of our work might be of interest to both ML researchers and to usability engineers who seek to improve the subjective satisfaction of users with websites.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Any design object needs effective presentation, in which structuring of textual and visual information is highly important <ref type="bibr" target="#b0">[1]</ref>. Many researchers and designers have been looking for the principles of harmonious organization of compositional elements in architecture and website design. For instance, visual appearance of web user interfaces (UIs) is known to affect behavior of users, and its analysis can help to improve usability and thus increase KPIs of the website, such as e.g. conversion rate <ref type="bibr" target="#b1">[2]</ref>. The visual complexity assessment helps to identify and describe problems in the website UI. Visual complexity is affected by the number of elements in an object or image, their structural relations, the detail of the information that these elements provide, etc. <ref type="bibr" target="#b2">[3]</ref>. It has been scientifically proven that aesthetic preferences for the visual complexity of web pages are influenced by users' age and previous experience <ref type="bibr" target="#b3">[4]</ref>. However, our article focuses on the dependence of visual complexity in web UI screenshots: namely the common compositional elements in web pages (buttons, texts, lists, etc.), as well as multimedia elements (images, videos, etc.).</p><p>The identification of UI elements in a website page screenshot for further assessment of visual complexity can be obtained through either manual labeling or automated recognition process <ref type="bibr" target="#b4">[5]</ref>. Automation of any process makes it possible to simplify it and helps to free a person from routine and tedious tasks, but often it involves additional costs and resources (time, labor, etc.), especially at the initial stage. Table <ref type="table" target="#tab_0">1</ref> shows a comparison of the automatic and manual methods with respect to UI labeling.</p><p>Thus, the purpose of the current work is to determine the types of elements that affect the subjective perception of websites, as well as to compare the models built with the factors' values obtained via automated vs. manual labeling of web UI screenshots. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">The Study Description 2.1. The Manual Labeling</head><p>In the experiment, the subjects were offered about 500 website interface screenshots and asked to label UI elements in them: highlight the element in a box and identify the elements' type. They were using a dedicated software tool, LabelImg (see in Fig. <ref type="figure" target="#fig_0">1</ref>). In total, 11 human labelers took part in this activity, after providing informed consent. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">The Automated Labeling</head><p>In addition to the manual labeling of the screenshots, we also performed their automated analysis, using our dedicated Visual Analyzer (VA) software tool, available at http://va.wuikb.info/ and described in detail in <ref type="bibr" target="#b5">[6]</ref>. It identifies UI elements in images based on previously trained ML models (see in Fig. <ref type="figure" target="#fig_1">2</ref>), but our previous studies suggest that its accuracy is somehow deficient, particularly in determining the type of each UI element. In Table <ref type="table" target="#tab_1">2</ref> we show the types of UI elements that had been identified in manual and automated labeling that we performed. As one can seen from it, the main difference between the two methods lies in determining the types of elements. In the manual labeling, the participants were able to rather successfully identify 26 different UI types, while for the automated labeling, there were 8 types, resulting from the pre-trained ML models. In addition, due to high visual diversity of today's web designs, the accuracy of the automated type detection was the most problematic dimension in the semantic-spatial analysis that our VA tool performed (see <ref type="bibr" target="#b5">[6]</ref> for more detail).</p><p>From the labeling datasets, the heights and widths of the elements were calculated (also by the elements' types: text paragraphs, buttons, images, panels, etc.), and there was also the data on the height and width of the screenshot. Based on this information, areas of each UI he element was calculated, as well as the share occupied by this element in the total screenshot space (by the elements' types: texts, images, background, etc.).</p><p>Thus, there were 2 groups of factors in the behavior models that we further constructed for complexity, aesthetics and orderliness that were the dependent variables in the study:</p><p>1) the number of elements: the number of elements of a certain type located in one screenshot; 2) the proportions: the shares of the total area of the elements' types to the total area of the screenshot. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.">The Subjective Perception Evaluation</head><p>For each screenshot, we also had evaluations of complexity, orderliness and aesthetics, provided by another 137 participants (67 female, 70 male). The majority of them were Russians (89.1%), while the rest were from Bulgaria, Germany, South Africa, etc. More details on the participants and the procedure can be found in one of our previous works <ref type="bibr" target="#b6">[7]</ref>.</p><p>Each of the subjective perception dimensions was assessed on a scale from 1 to 10. Then the average values for each indicator were found using formulas (1)-( <ref type="formula" target="#formula_2">3</ref>), where i is the number of web pages, n j is the number of participants who provided the evaluations for the i-th website.</p><p>∑</p><p>where indicator of the aesthetics of the interface of the i-th web page, estimation of the aesthetics of the interface of the i-th web page ∑</p><p>where indicator of the complexity of the interface of the i-th web page, estimation of the complexity of the interface of the i-th web page ∑</p><p>where indicator of the ordering of the interface of the i-th web page, estimation of the ordering of the interface of the i-th web page.</p><p>Complexity is the number of elements on the screen and their arrangement. Orderliness is an ordered data structure of the web interface that allows the user to easily find the information they need. Aesthetics is an assessment of the attractiveness of a product by the user. This indicator is important because the aesthetics of any web interface has a strong impact on the user, even when he tries to evaluate the functionality of the system <ref type="bibr" target="#b7">[8]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Results</head><p>Using the SPSS statistical analysis software, we build linear regression user behavior models (as representing the most universal ML method) with the factors resulting from the manual and the automated labeling. The dependent variables in the models were the three subjective perception evaluation scores, while the independent variables were the factors resulting from the two labeling processes. The step-by-step method of the analysis allowed stepwise inclusion of the factors in the models, thereby discarding those that did not make a significant contribution to explaining the dependent variables.</p><p>The results of the regression analysis are presented in Tables <ref type="table" target="#tab_2">3-5</ref>, each corresponding to a different subjective perception dimension. The null hypothesis H 0 was as usual in the regression analysis, that the regression equation is not significant. For each of the dependent variable, the models turned out to be significant (p &lt; 0.05), while the significances for the factors selected by the step-by-step method are shown in the respective columns of the tables. As one can note from Table <ref type="table" target="#tab_2">3</ref>, the perceived complexity of websites was most influenced by the amount of text, the number of radio buttons, the number of buttons and tabs. The subjective orderliness (Table <ref type="table" target="#tab_3">4</ref>) was most influenced by the proportion of radio buttons and buttons. The perception of aesthetics (Table <ref type="table">5</ref>) was most influenced by the proportion and amount of text on the page, the number of buttons, labels and images. Despite the fact that the constructed models had rather low determination coefficients (R 2 ), it is still possible to draw a general conclusion about which elements affect the assessment of website perception among users, and whether the degree of this dependence is influenced by the labeling method (Table <ref type="table">6</ref>).</p><p>As one can see from Table <ref type="table">6</ref>, the share of text on the page was significant for all the three subjective perception dimensions. Moreover, the presence of the text had the greatest impact on the assessment of complexity and aesthetics. At the same time, the obtained values for the automated and the manual labeling did not differ much (4-10%), but with the automated labeling, the models' quality indexes, as represented by R 2 s, were somehow superior. the share of the text area, text style (font, size, etc.), and the presence of pictures that dilute the text, and many other factors are also important. The number of images, their size, quality and position on the web page affect the overall website subjective perception, including the assessment of the aesthetics and orderliness. As a general rule, these indicators are interchangeable: the lack of text is often compensated by a variety of graphic elements and images. Therefore, in order for web UI to comply with usability standards and to be simpler and more understandable for its users, it is necessary not to clutter the interface with a large number of elements and break long texts into smaller parts. At the same time, the number of element types did not significantly affect any of the subjective perception dimensions.</p><p>The quality of the user behavior models built on the factors resulting from the automated labeling was slightly higher than for the manual one, which suggests feasibility of our UI visual analysis tool. In general, this may indicate that the automation of the labeling process makes sense, but it requires high costs and the presence of certain knowledge to implement it. After all, our VA software was initially trained with the data once provided by human labelers <ref type="bibr" target="#b5">[6]</ref>.</p><p>It should be noted that the constructed linear models have low R 2 coefficients, so they should not be used in production. The goal of our current study was merely to compare the two labeling methods, while for real user behavior modeling, more advanced ML methods and architectures should be used.</p><p>Our plans for further research include investigation of the effects of web page layouts (the size of the element, its type and occupied area) on the subjective assessment of the perception of the site, but also the colors, fonts, types of buttons, animations, etc.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Manual UI labeling with LabelImg software tool (the selection of UI type is at the bottom)</figDesc><graphic coords="2,118.00,432.20,360.47,310.25" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Automated UI labeling with our Visual Analyzer tool (an example)</figDesc><graphic coords="3,93.85,178.98,422.44,336.30" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Advantages and disadvantages of automated and manual methods</figDesc><table><row><cell>The labeling method</cell><cell>Advantages</cell><cell>Disadvantages</cell></row><row><cell>Manual</cell><cell>1) Higher accuracy in</cell><cell>1) It takes a large number of people</cell></row><row><cell></cell><cell>determining a UI element's</cell><cell>to conduct the analysis</cell></row><row><cell></cell><cell>type</cell><cell>2) Human factor: grammatical errors,</cell></row><row><cell></cell><cell>2) More robust list of</cell><cell>incorrect definition of the type of</cell></row><row><cell></cell><cell>various UI elements' types</cell><cell>element, inconsistency in</cell></row><row><cell></cell><cell>can be used</cell><cell>understanding, etc., which can distort</cell></row><row><cell></cell><cell></cell><cell>the final result</cell></row><row><cell>Automated</cell><cell>1) The analysis process</cell><cell>1) Image recognition is still</cell></row><row><cell></cell><cell>takes less time</cell><cell>computationally expensive and</cell></row><row><cell></cell><cell>2) No need to involve paid</cell><cell>inaccurate in some aspects</cell></row><row><cell></cell><cell>labelers</cell><cell>2) The cost of implementing the code</cell></row><row><cell></cell><cell></cell><cell>and its debugging</cell></row><row><cell></cell><cell></cell><cell>3) The need for training data, e.g. for</cell></row><row><cell></cell><cell></cell><cell>detecting the UI elements' types</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2</head><label>2</label><figDesc>The UI elements' types used in the manual and the automated labeling of web UIs</figDesc><table><row><cell>Element type</cell><cell>Manual labeling</cell><cell>Automated labeling</cell></row><row><cell>Button</cell><cell>+</cell><cell>+</cell></row><row><cell>Text</cell><cell>+</cell><cell>+</cell></row><row><cell>Checkbox</cell><cell>+</cell><cell>+</cell></row><row><cell>Radio button</cell><cell>+</cell><cell>+</cell></row><row><cell>Dropdown</cell><cell>+</cell><cell>+</cell></row><row><cell>Label</cell><cell>+</cell><cell>+</cell></row><row><cell>Image</cell><cell>+</cell><cell>+</cell></row><row><cell>Link</cell><cell>+</cell><cell>+</cell></row><row><cell>Background image</cell><cell>+</cell><cell>-</cell></row><row><cell>Panel</cell><cell>+</cell><cell>-</cell></row><row><cell>Tabs</cell><cell>+</cell><cell>-</cell></row><row><cell>Paragraph</cell><cell>+</cell><cell>-</cell></row><row><cell>Imagelink</cell><cell>+</cell><cell>-</cell></row><row><cell>Textinput</cell><cell>+</cell><cell>-</cell></row><row><cell>Textblock</cell><cell>+</cell><cell>-</cell></row><row><cell>Buttonimage</cell><cell>+</cell><cell>-</cell></row><row><cell>Headline</cell><cell>+</cell><cell>-</cell></row><row><cell>Image_background</cell><cell>+</cell><cell>-</cell></row><row><cell>Symbol</cell><cell>+</cell><cell>-</cell></row><row><cell>Back</cell><cell>+</cell><cell>-</cell></row><row><cell>Selectbox</cell><cell>+</cell><cell>-</cell></row><row><cell>Scroll bar</cell><cell>+</cell><cell>-</cell></row><row><cell>Date</cell><cell>+</cell><cell>-</cell></row><row><cell>List</cell><cell>+</cell><cell>-</cell></row><row><cell>Pagination</cell><cell>+</cell><cell>-</cell></row><row><cell>Table</cell><cell>+</cell><cell>-</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3</head><label>3</label><figDesc>The regression analysis results for the complexity dependent variable</figDesc><table><row><cell>Method type</cell><cell>Factor</cell><cell>R 2</cell><cell>Non-</cell><cell>Standard</cell><cell>t</cell><cell>Significance</cell></row><row><cell></cell><cell></cell><cell></cell><cell>standard</cell><cell>coef.</cell><cell></cell><cell></cell></row><row><cell></cell><cell></cell><cell></cell><cell>coef.</cell><cell></cell><cell></cell><cell></cell></row><row><cell>Automated</cell><cell>Count of</cell><cell>0.092</cell><cell>0.001</cell><cell>0.218</cell><cell>4.576</cell><cell>0.000</cell></row><row><cell></cell><cell>labels</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell>Count of text</cell><cell>0.127</cell><cell>0.012</cell><cell>0.195</cell><cell>4.131</cell><cell>0.000</cell></row><row><cell></cell><cell>Count of</cell><cell>0.136</cell><cell>0.039</cell><cell>0.096</cell><cell>2.137</cell><cell>0.033</cell></row><row><cell></cell><cell>radio buttons</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>Manual</cell><cell>Number of</cell><cell>0.062</cell><cell>0.001</cell><cell>0.172</cell><cell>3.427</cell><cell>0.001</cell></row><row><cell></cell><cell>elements on</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell>the page</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell>Share of</cell><cell>0.082</cell><cell>0.005</cell><cell>0.153</cell><cell>3.156</cell><cell>0.002</cell></row><row><cell></cell><cell>images</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell>Share of</cell><cell>0.095</cell><cell>0.002</cell><cell>-0.118</cell><cell>-2.699</cell><cell>0.007</cell></row><row><cell></cell><cell>panels</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell>Share of</cell><cell>0.107</cell><cell>0.017</cell><cell>0.119</cell><cell>2.790</cell><cell>0.005</cell></row><row><cell></cell><cell>buttons</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell>Share of</cell><cell>0.114</cell><cell>0.007</cell><cell>0.089</cell><cell>2.042</cell><cell>0.042</cell></row><row><cell></cell><cell>paragraphs</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell>Share of tabs</cell><cell>0.121</cell><cell>0.032</cell><cell>0.087</cell><cell>2.025</cell><cell>0.043</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4</head><label>4</label><figDesc>The regression analysis results for the orderliness dependent variable</figDesc><table><row><cell>Method type</cell><cell>Factor</cell><cell>R 2</cell><cell>Non-</cell><cell>Standard</cell><cell>t</cell><cell>Significance</cell></row><row><cell></cell><cell></cell><cell></cell><cell>standard</cell><cell>coef.</cell><cell></cell><cell></cell></row><row><cell></cell><cell></cell><cell></cell><cell>coef.</cell><cell></cell><cell></cell><cell></cell></row><row><cell>Automated</cell><cell>Share of text</cell><cell>0.027</cell><cell>1.221</cell><cell>-0.162</cell><cell>-3.604</cell><cell>0.000</cell></row><row><cell></cell><cell>Screenshot</cell><cell>0.048</cell><cell>0.000</cell><cell>.253</cell><cell>4.793</cell><cell>0.000</cell></row><row><cell></cell><cell>width</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell>Count of labels</cell><cell>0.080</cell><cell>0.001</cell><cell>-0.199</cell><cell>-3.763</cell><cell>0.000</cell></row><row><cell></cell><cell>Count of</cell><cell>0.096</cell><cell>0.005</cell><cell>-0.123</cell><cell>-2.722</cell><cell>0.007</cell></row><row><cell></cell><cell>buttons</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell>Share of radio</cell><cell>0.107</cell><cell>14.489</cell><cell>-0.104</cell><cell>-2.305</cell><cell>0.022</cell></row><row><cell></cell><cell>buttons</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>Manual</cell><cell>Share of text</cell><cell>0.024</cell><cell>0.004</cell><cell>-0.179</cell><cell>-4.104</cell><cell>0.000</cell></row><row><cell></cell><cell>Share of</cell><cell>0.051</cell><cell>0.001</cell><cell>0.138</cell><cell>3.115</cell><cell>0.002</cell></row><row><cell></cell><cell>background</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell>images</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell>Share of links</cell><cell>0.067</cell><cell>0.011</cell><cell>-0.125</cell><cell>-2.828</cell><cell>0.005</cell></row><row><cell></cell><cell>Share of</cell><cell>0.077</cell><cell>0.057</cell><cell>0.103</cell><cell>2.354</cell><cell>0.019</cell></row><row><cell></cell><cell>"back" arrows</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>Table 5</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="5">The regression analysis results for the aesthetics dependent variable</cell><cell></cell><cell></cell></row><row><cell>Method type</cell><cell>Factor</cell><cell>R 2</cell><cell>Non-</cell><cell>Standard</cell><cell>t</cell><cell>Significance</cell></row><row><cell></cell><cell></cell><cell></cell><cell>standard</cell><cell>coef.</cell><cell></cell><cell></cell></row><row><cell></cell><cell></cell><cell></cell><cell>coef.</cell><cell></cell><cell></cell><cell></cell></row><row><cell>Automated</cell><cell>Screenshot</cell><cell>0.072</cell><cell>0.000</cell><cell>0.360</cell><cell>7.042</cell><cell>0.000</cell></row><row><cell></cell><cell>width</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell>Share of text</cell><cell>0.119</cell><cell>1.549</cell><cell>-0.207</cell><cell>-4.749</cell><cell>0.000</cell></row><row><cell></cell><cell>Count of</cell><cell>0.144</cell><cell>0.007</cell><cell>-0.147</cell><cell>-3.373</cell><cell>0.001</cell></row><row><cell></cell><cell>button</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell>Count of label</cell><cell>0.161</cell><cell>0.002</cell><cell>-0.155</cell><cell>-3.032</cell><cell>0.003</cell></row><row><cell>Manual</cell><cell>Share of</cell><cell>0.046</cell><cell>0.002</cell><cell>0.224</cell><cell>4.933</cell><cell>0.000</cell></row><row><cell></cell><cell>background</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell>images</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell>Share of text</cell><cell>0.079</cell><cell>0.007</cell><cell>-0.281</cell><cell>-5.178</cell><cell>0.000</cell></row><row><cell></cell><cell>Share of links</cell><cell>0.099</cell><cell>0.014</cell><cell>-0.123</cell><cell>-2.823</cell><cell>0.005</cell></row><row><cell></cell><cell>Count of text</cell><cell>0.114</cell><cell>0.014</cell><cell>0.152</cell><cell>2.786</cell><cell>0.006</cell></row><row><cell></cell><cell>Count of</cell><cell>0.122</cell><cell>0.003</cell><cell>0.095</cell><cell>2.114</cell><cell>0.035</cell></row><row><cell></cell><cell>images</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Acknowledgment</head><p>The reported study was funded by RFBR according to the research project No. 19-29-01017.</p></div>
			</div>

			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>To improve the quality of the constructed user behavior models, we tried reducing the number of factors and took only the common ones for the two methods, namely:</p><p>1. The total number of elements in the interface 2. The number of types of elements on the page 3. Percentage of space under the text 4. Percentage of space under images 5. Average height of elements 6. Average width of elements 7.</p><p>The average size was taken into account for elements such as labels, buttons, test blocks and images. For text and images, a size restriction has been imposed: no more than 300 px in height and no more than 1000 px in width. The results of the regression analysis for the three dimensions of the subjective perception are shown in Table <ref type="table">7</ref>. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Conclusion</head><p>The study showed that the subjective assessment of the website perception is influenced by the amount of text on the page and the share of images: the more text on the page, the more complex and less aesthetic the website appears. However, this statement is not entirely correct, since in addition to</p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Assessment of visual perception of information design objects</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">P</forename><surname>Rassadina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><forename type="middle">V</forename><surname>Ivanova</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. XVI Int Conf on Cultural studies, philology, art history: urgent problems of modern science</title>
				<meeting>XVI Int Conf on Cultural studies, philology, art history: urgent problems of modern science</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="30" to="35" />
		</imprint>
	</monogr>
	<note>-In Russian</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Influence of behavioral factors of ranking on the position of sites in search results</title>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">V</forename><surname>Pesterev</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Bulletin of the Belgorod State Technological University</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="219" to="221" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note>-In Russian</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Endorsement and visual complexity in food advertising on Instagram</title>
		<author>
			<persName><forename type="first">S</forename><surname>Kusumasondjaja</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Tjiptono</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Internet Research</title>
		<imprint>
			<biblScope unit="volume">29</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="659" to="687" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">An investigation into visual complexity and aesthetic preference to facilitate the creation of more appropriate learning analytics systems for children</title>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">F</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">H</forename><surname>Lin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computers Human Behavior</title>
		<imprint>
			<biblScope unit="volume">92</biblScope>
			<biblScope unit="page" from="706" to="715" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">System and method for selecting significant page elements with implicit indication of coordinates for identifying and viewing relevant information</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">V</forename><surname>Tsyplyaev</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">A</forename><surname>Vinokurov</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Patent</title>
		<imprint>
			<biblScope unit="volume">2708790</biblScope>
			<biblScope unit="page">C2</biblScope>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Auto-extraction and integration of metrics for web user interfaces</title>
		<author>
			<persName><forename type="first">M</forename><surname>Bakaev</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Heil</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Khvorostov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gaedke</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Web Engineering</title>
		<imprint>
			<biblScope unit="volume">17</biblScope>
			<biblScope unit="issue">6&amp;7</biblScope>
			<biblScope unit="page" from="561" to="590" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">I Don&apos;t Have That Much Data! Reusing User Behavior Models for Websites from Different Domains</title>
		<author>
			<persName><forename type="first">M</forename><surname>Bakaev</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Speicher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Heil</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gaedke</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. 20th International Conference on Web Engineering (ICWE 2020)</title>
				<meeting>20th International Conference on Web Engineering (ICWE 2020)<address><addrLine>Cham</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="146" to="162" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Developing emotional design: Emotions as cognitive processes and their role in the design of interactive technologies</title>
		<author>
			<persName><forename type="first">S</forename><surname>Triberti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Chirico</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>La Rocca</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Riva</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Frontiers in psychology</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="page">1773</biblScope>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
