<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Ordinal Scale Evaluation of Smiling Intensity using Comparison-Based Network</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Kei</forename><surname>Shimonishi</surname></persName>
							<email>shimonishi@i.kyoto-u.ac.jp</email>
							<affiliation key="aff0">
								<orgName type="institution">Kyoto University</orgName>
								<address>
									<addrLine>Yoshida-honmachi</addrLine>
									<settlement>Sakyo</settlement>
									<region>Kyoto</region>
									<country key="JP">Japan</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Kazuaki</forename><surname>Kondo</surname></persName>
							<email>kondo@ccm.media.kyoto-u.ac.jp</email>
							<affiliation key="aff0">
								<orgName type="institution">Kyoto University</orgName>
								<address>
									<addrLine>Yoshida-honmachi</addrLine>
									<settlement>Sakyo</settlement>
									<region>Kyoto</region>
									<country key="JP">Japan</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Hirotada</forename><surname>Ueda</surname></persName>
							<email>ueda.hirotada.2r@kyoto-u.ac.jp</email>
							<affiliation key="aff0">
								<orgName type="institution">Kyoto University</orgName>
								<address>
									<addrLine>Yoshida-honmachi</addrLine>
									<settlement>Sakyo</settlement>
									<region>Kyoto</region>
									<country key="JP">Japan</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Yuichi</forename><surname>Nakamura</surname></persName>
							<email>yuichi@media.kyoto-u.ac.jp</email>
							<affiliation key="aff0">
								<orgName type="institution">Kyoto University</orgName>
								<address>
									<addrLine>Yoshida-honmachi</addrLine>
									<settlement>Sakyo</settlement>
									<region>Kyoto</region>
									<country key="JP">Japan</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff1">
								<orgName type="department">Machine Learning for Cognitive and Mental Health Workshop (ML4CMH)</orgName>
								<address>
									<addrLine>AAAI 2024</addrLine>
									<settlement>Vancouver</settlement>
									<region>BC</region>
									<country key="CA">Canada</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Ordinal Scale Evaluation of Smiling Intensity using Comparison-Based Network</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">35332EB50D7A75CF5E039C8A0EB1B75B</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T16:45+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Facial expression recognition</term>
					<term>Siamese network</term>
					<term>ranking</term>
					<term>ordinal scales</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The ability to evaluate both explicit facial expressions and intermediate expressions is helpful for human monitoring. Since intermediate facial expression is out of the scope of traditional studies, evaluation scores obtained from traditional facial expression recognition techniques are unreliable. In this paper, we propose an ordinal scale-based evaluation scheme for facial expression based on a comparison. The proposed framework is based on an ordinal scale; it is challenging to construct a standard scale that can be applied to multiple individuals. However, it is expected to be effective enough to track changes in the facial expressions of the specific individual, including intermediate expressions. We also propose an algorithm for selecting reference images from the data by taking into account the consistencies of the strong-weak relationships between reference images because the reference image selection significantly impacts the ordinal evaluation. Our approach is evaluated by conducting experiments with human annotators.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Monitoring an individual's Quality of Life (QOL) is becoming increasingly important to maintain good mental conditions and detect early trends in harmful conditions. Because direct QOL inquires are bothering and it is difficult to accurately represent one's internal state, estimating internal state from external nonverbal information is desired. Facial expression is one of the modalities that reflects an individual's internal state and is expressed with being influenced by mental condition. For example, when an individual is not feeling well, the same smile may appear weaker than usual. Therefore, monitoring facial expressions in daily life is a crucial clue to estimating an individual's QOL.</p><p>The research field of facial expression recognition (FER) has a long history, and it has already been put into practical use as a technology, such as smiling shutters. While traditional FER mainly focuses on recognizing whether a clear facial expression is represented or not, from the viewpoint of monitoring in daily life, evaluating the degree of expression for the individual is rather crucial, especially for patients with dementia who have little or no facial expressions. Based on this point of view, this research aims to draw a curve of transitions of the individual's degree of facial expressions, particularly smiling intensity, as shown in Figure <ref type="figure" target="#fig_5">1</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Figure 1: An example of transition curve of smiling intensity in daily life</head><p>Though the traditional algorithm of FER seems able to evaluate intermediate facial expressions as a probability that a specific facial expression is represented, the probability values are not so reliable, especially for evaluating intermediate expressions. This is because the intermediate facial expression was out of the scope of the traditional studies; learning is likely to output a value close to the binary value of either no expression (0) or an expression <ref type="bibr" target="#b0">(1)</ref>. As a result, for example, when the degree of smile expression is estimated for a series of facial expressions, the value may change abruptly over a series of times as shown in Figure <ref type="figure">2</ref>. In addition, it is also difficult for the machine learning algorithm to directly learn the intermediate facial expressions since it is difficult for even humans to give appropriate absolute values for intermediate facial expressions.</p><p>Kondo et al. <ref type="bibr" target="#b0">[1]</ref> proposed a network for recognizing smiling based on "comparison" to address the issues of recognizing intermediate facial expressions. Their work By borrowing this comparison-based idea to evaluate facial expressions, we propose an approach to evaluate smiling intensity with an ordinal scale. The basic idea of this approach is that if we have multiple reference face images for a specific individual and a method for comparing facial expressions, we can evaluate the smiling intensity of a new image of the individual through pairwise comparison with the reference images, as shown in Figure <ref type="figure">3</ref>.</p><p>Since the expression ratings in this method are based on an ordinal scale, the degree of each rating is not the same for multiple individuals. However, this ordinal scale-based approach may satisfy our need to capture changes in facial expressions for each individual.</p><p>In addition, reference image selection is crucial for this ordinal-based evaluation because they are considered an evaluation space for facial expressions. Therefore, we also propose an algorithm of reference image selection from a large number of face image data of each individual based on consistencies of comparison results within images.</p><p>In summary, the contributions of this paper are as follows:</p><p>• We propose an approach to evaluating intermediate smiling intensity by ordinal scales based on comparisons. • We propose an algorithm for selecting appropriate reference images to construct appropriate evaluation space.</p><p>We briefly introduce related work in the next section. Then, we introduce an approach to evaluate facial expressions by ordinal scales and an algorithm of reference image selection. We evaluate our approach and algorithm with human annotators, and finally, we conclude our research.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Work</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Facial expression recognition</head><p>Facial expression recognition is widely utilized in several fields. Traditional studies mainly focused on determining whether a specific expression is represented or not.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.1.">Facial Action Coding Systems</head><p>Facial Action Coding Systems (FACS) <ref type="bibr" target="#b1">[2]</ref> is a framework proposed by Ekman et al. that classifies a face into several parts (Action Units; AUs) based on the basic action units of individual muscles and describes facial expressions as a combination of these AU actions. Many facial expression recognition applications have used FACS as features, and for example, OpenFace <ref type="bibr" target="#b2">[3]</ref> can analyze multiple facial expressions in near real-time by automatically recognizing the actions of AUs.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.2.">Deep neural network based approach</head><p>Although the FACS-based FER approach has been successful, it has the limitation that the final results are affected by the accuracy of FACS detection. This limitation can become a problem, especially when trying to capture subtle differences in facial expressions because the effect of observation noise cannot be ignored. On the other hand, the end-to-end approach by the deep neural network can be expected to reduce the effect of such observation noise by eliminating the necessity of explicit feature detection. For example, VGGNet <ref type="bibr" target="#b3">[4]</ref> is a traditional deep neural network, but it is known that human facial features can be extracted well, and recent research of FER also utilized VGGNet <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b5">6]</ref> </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Siamese structure-based recognition technique</head><p>Siamese network <ref type="bibr" target="#b6">[7]</ref> is one of the deep neural networks of metric learning. This network acquires two inputs and returns the distance between the two inputs. By applying the same structures and the same weights to feature extraction layers of these two inputs and the distance of the two inputs to the loss function, the network can learn the distance space. The Siamese Network is a network that determines whether two inputs are similar or different and has been applied to handwritten signature  recognition <ref type="bibr" target="#b7">[8]</ref> and used as a framework for anomaly detection <ref type="bibr" target="#b8">[9]</ref>. As one of its features, it is known as a network that can be trained from a small number of training data compared to conventional networks that perform multi-valued discrimination and regression <ref type="bibr" target="#b9">[10]</ref>. Kondo et al. <ref type="bibr" target="#b0">[1]</ref> proposed an approach to the evaluation of facial expressions based on comparison inspired by the Siamese structure. Their approach compares two facial images and returns which of one image represents more smiles, and they showed that the approach has the potential to distinguish subtle facial expression differences. In addition, Zhang et al. <ref type="bibr" target="#b10">[11]</ref> extended their work from a positive-neutral direction to a negative-neutral direction.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Comparison-based smiling evaluation by ordinal scales</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Overview of the proposed framework</head><p>As introduced in the Introduction, the basic idea of our approach is a comparison-based evaluation. Kondo et al. <ref type="bibr" target="#b0">[1]</ref> has developed a Siamese-based smiling recognition network that takes two face images as input and recognizes which one is expressing smiling more. By borrowing this idea, once we develop a network that can determine which of two images represents more smiles, and if we have multiple reference images, we can evaluate the smile intensity of a new image through pairwise comparison with the reference images as also introduced in the Introduction. When it comes to determining smiling intensity based on ordinal scales, although all the comparison results are ideally consistent, the results are sometimes inconsistent due to an ambiguity of slightly different face images. Therefore, we apply a voting-based evaluation and determine smiling scores by merging multiple comparison results. In addition, we propose an algorithm to select appropriate reference images to reduce the ambiguity between reference images in the following section.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">A network for facial expression comparison</head><p>In this paper, we defined the recognition task as a simple two-category classification problem (i.e., determining which of two input images represents the greater degree of smiling) and construct a Siamese-based network to recognize smiling similar to the network Kondo et al.</p><p>have developed <ref type="bibr" target="#b0">[1]</ref>.</p><p>Figure <ref type="figure" target="#fig_2">4</ref> shows the structure of the proposed network that accepts two input images and returns two likelihood values corresponding to ascension and descension labels relative to the degree of smiling. We employed the CNN component of VGG16 <ref type="bibr" target="#b3">[4]</ref> and two fully connected layers with rectified linear units, a 0.25 dropout rate, and SoftMax in the proposed method. The ground-truth likelihood values for an input image pair were represented as a two-element one-hot vector, with its element corresponding to the ground truth label set to 1 and the other element set to 0, respectively. We used categorical cross-entropy loss to optimize the network parameters, as follows:</p><formula xml:id="formula_0">𝐿𝑐𝑎𝑡 = − ∑︁ 𝑖 {𝑦𝑖𝑙𝑜𝑔(𝑦 ˆ𝑖) + (1 − 𝑦𝑖)𝑙𝑜𝑔(1 − 𝑦 ˆ𝑖)} ,<label>(1)</label></formula><p>where 𝑖 = {0, 1}, 𝑦𝑖, and 𝑦 ˆ𝑖 denotes ascension and descension labels relative to the degree of smiling, the ground-truth label, and the predicted likelihood values, respectively.</p><p>The previously proposed network by Kondo et al. was not designed to consider the order of inputs, resulting in instances where swapping the order of two inputs led to contradictory outputs. To address this issue, we input a permuted version of the two features extracted from two input images by the CNN component into the fully connected layer in the latter stage and calculate the categorical cross-entropy loss of inverted input, 𝐿𝑖𝑛𝑣, as same as 𝐿𝑐𝑎𝑡, as shown in red arrows in Figure <ref type="figure" target="#fig_2">4</ref>. Also, a loss of consistency of these two types of input is calculated as where 𝑃 𝐴𝑠 (𝑛, 𝑚) and 𝑃 𝐷𝑒𝑠 (𝑛, 𝑚) represent probabilities that the degree of smiling of image 𝐼𝑛 is larger or smaller than that of image 𝐼𝑚, respectively. In other words, 𝑃 (𝑠𝑛(𝐼𝑛) &gt; 𝑠𝑛(𝐼𝑚)) and 𝑃 (𝑠𝑛(𝐼𝑛) &lt; 𝑠𝑛(𝐼𝑚)), where 𝑠𝑛(𝐼) represents a degree of smiling of image 𝐼. Also, 𝑃 𝑓 and 𝑃𝑖 represent the likelihoods of the forward comparison stream and inverse comparison stream, respectively. In total, our network is trained to decrease the following loss function:</p><formula xml:id="formula_1">𝐿𝑐𝑜𝑛 = 1 − {𝑚𝑖𝑛(𝑃 𝐴𝑠 𝑓 (𝑡𝑔𝑡, 𝑟𝑒𝑓 ), 𝑃 𝐴𝑠 𝑖 (𝑡𝑔𝑡, 𝑟𝑒𝑓 ))</formula><formula xml:id="formula_2">𝐿 = 𝐿𝑐𝑎𝑡 + 𝐿𝑖𝑛𝑣 + 𝐿𝑐𝑜𝑛.<label>(3)</label></formula><p>Here, we expected that the CNN and the fully connected components would be trained to compare extracted features in order to project the results onto the likelihood values of the ascension and descension labels, respectively.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">Voting-based evaluation</head><p>Since the reference images may include some ambiguity between neighboring images, it is difficult to directly determine the degree of smiling of the new target image in a reference image set. Therefore, we apply a voting technique to determine the final rank of the image. The algorithm votes to possible ranks using the result of each comparison of reference images and a target image. As a result, the most likely rank should have a maximum number of votes.</p><p>In particular, the procedure is as follows. Suppose that we have 𝑁 reference images with its order of degree of smiling, i.e., 𝑠𝑛(𝐼𝑖) &gt; 𝑠𝑛(𝐼𝑗), ∀𝑖 &lt; 𝑗. A new target image 𝐼𝑛𝑒𝑤 is compared to all reference images, and likelihood values that the degree of smiling of target image is larger than that of a reference image 𝑃 𝐴𝑠 (𝑛𝑒𝑤, 𝑛) and likelihood that the degree of smiling of target image is lower than that of a reference image 𝑃 𝐷𝑒𝑠 (𝑛𝑒𝑤, 𝑛) for all reference images (𝑛 ∈ {1, . . . , 𝑁 }) are obtained. Because if 𝑠𝑛(𝐼𝑛𝑒𝑤) &lt; 𝑠𝑛(𝐼𝑛) is estimated, the smile rank of 𝐼𝑛𝑒𝑤 is estimated as larger than 𝑛, large values are voted to ranks larger than 𝑛. In practice, add likelihood values of "ascend" and "descend" to the ranks lower than and higher than 𝑛, respectively.</p><p>Simply thinking, the degree of smiling in the reference images can be determined by searching the position whose scores are maximum. Here, the position 𝑟 can be derived as:</p><formula xml:id="formula_3">𝑟 = arg max 𝑟 {︃ 𝑟−1 ∑︁ 𝑛=1 𝑃 𝐷𝑒𝑠 (𝑛𝑒𝑤, 𝑛) + 𝑁 ∑︁ 𝑛=𝑟+1 𝑃 𝐴𝑠 (𝑛𝑒𝑤, 𝑛)</formula><p>}︃ .</p><p>(4) In addition, we here apply a mean-shift algorithm to determine the evaluation score based on these probability values.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Reference image selection</head><p>To apply voting-based ordinal-scale evaluation as described above, we first need to construct an evaluation space with several reference images. Since the proposed approach utilizes ordinal scales, the construction of the evaluation space is crucial for the capability of the approach. Although a straightforward way is to utilize all the face data as reference images, the evaluation space constructed by very similar or subtle different images is unreliable due to the ambiguity of these images.</p><p>In this paper, we first consider all the data as baseline images and take pair-wise comparisons to sort all the data in a dataset and construct a baseline ranking. Then, we select several images from the baseline ranking as reference images and quantize the evaluation space by taking into account consistency to address the issues due to ambiguity.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Baseline ranking construction</head><p>Figure <ref type="figure">6</ref> (a) shows a comparison table of the result of all pair-wise comparisons in baseline images. Each color shows a probability of how a target image has a stronger smile than a reference image, i.e., 𝑃 𝐴𝑠 (𝑡𝑔𝑡, 𝑟𝑒𝑓 ). The blue shows a pair whose target image has a stronger smile than the reference image, i.e., 𝑃 𝐴𝑠 (𝑡𝑔𝑡, 𝑟𝑒𝑓 ) &gt; 𝑃 𝐷𝑒𝑠 (𝑡𝑔𝑡, 𝑟𝑒𝑓 ). In contrast, the red area shows a pair whose reference image has a stronger smile than a target image, i.e., 𝑃 𝐴𝑠 (𝑡𝑔𝑡, 𝑟𝑒𝑓 ) &lt; 𝑃 𝐷𝑒𝑠 (𝑡𝑔𝑡, 𝑟𝑒𝑓 ). The white area represents that the target and the reference images represent similar facial expressions.</p><p>By sorting the baseline images based on the sum of the probability values in each column of this table, a baseline ranking considering the consistency of the strong-weak relationship can be constructed (Figure <ref type="figure">6</ref>  </p><formula xml:id="formula_4">𝐿 = 𝑁 ∑︁ 𝑛 ⎧ ⎪ ⎨ ⎪ ⎩ ∑︁ {𝑛 ^|𝐼 𝑛 ^∈{𝐼 𝑟𝑎𝑛𝑘 1 ,...,𝐼 𝑟𝑎𝑛𝑘 𝑛−1 }} 𝑃𝐷𝑒𝑠(𝑛, 𝑛 ˆ) + ∑︁ {𝑛 ^|𝐼 𝑛 ^∈{𝐼 𝑟𝑎𝑛𝑘 𝑛+1 ,...,𝐼 𝑟𝑎𝑛𝑘 𝑁 }} 𝑃𝐴𝑠(𝑛, 𝑛 ˆ)⎫ ⎪ ⎬ ⎪ ⎭<label>(5)</label></formula><p>By maximizing this total consistency, baseline ranking images (𝐼 𝑟𝑎𝑛𝑘 𝐿 can be obtained. From now on, the subscript 𝑛 will be used to sort the images in descending order of smiling degree.</p><p>An example of the consistency of the strong-weak relations in this rearranged table is shown in Figure <ref type="figure">6</ref> (c) by replacing 𝑃 𝐴𝑠 (𝑡𝑔𝑡, 𝑟𝑒𝑓 ) into 𝑃 𝐷𝑒𝑠 (𝑡𝑔𝑡, 𝑟𝑒𝑓 ) when 𝑠𝑛(𝐼𝑡𝑔𝑡) &lt; 𝑠𝑛(𝐼 𝑟𝑒𝑓 ). We here denote probabilities 𝐶 𝑡𝑔𝑡,𝑟𝑒𝑓 to indicate this consistency as follows:</p><formula xml:id="formula_5">𝐶 𝑡𝑔𝑡,𝑟𝑒𝑓 = {︃ 𝑃 𝐴𝑠 (𝑡𝑔𝑡, 𝑟𝑒𝑓 ) if 𝑠𝑛(𝑡𝑔𝑡) &gt; 𝑠𝑛(𝑟𝑒𝑓 ), 𝑃 𝐷𝑒𝑠 (𝑡𝑔𝑡, 𝑟𝑒𝑓 ) if 𝑠𝑛(𝑡𝑔𝑡) &lt; 𝑠𝑛(𝑟𝑒𝑓 ).<label>(6)</label></formula><p>Ideally, all cells would be blue, i.e., consistency is nearly equal to 1. However, due to the ambiguity of comparison results for similar facial expressions, there is also ambiguity in the consistency between the neighboring baseline ranking. Therefore, reference image selection is important to construct appropriate evaluation space.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Reference image selection</head><p>An important factor in selecting reference images is the consistency of the strong-weak relationships within reference images. That is, when the consistency table is  calculated the same as the bottom-right figure of Figure <ref type="figure">6</ref>, the less red and white colors area is a better sign of reference image selection.</p><p>To realize that, we focus on a square region of neighbor images as shown in Figure <ref type="figure" target="#fig_7">7</ref>, and call this square consistency square. Suppose the differences between images are significant, and a strong-weak relationship is evident in the images. In that case, the consistency values in the consistency square are also expected to be large, i.e., cells in the square become blue. In contrast, when images are similar, and therefore the difference between images is ambiguous, the consistency values in the consistency square become low, i.e., cells in the square become white and red.</p><p>The basic idea of building a consistent evaluation space is to quantize images with low consistency values in the consistency square into a single class. As a result, the ambiguity between these images becomes "don't care" in the evaluation space, and the consistency of the evaluated value becomes significant. That is, it is good to select images where the total consistency values in the sum of consistency squares, as shown in the right of Figure <ref type="figure" target="#fig_7">7</ref>, becomes low.</p><p>In addition, neighbor images in the baseline ranking should not be selected as reference images. In other words, to select good reference images, the evaluation space is better to be divided evenly. To realize that, select a group of reference images so that the sum of the area of consistency space becomes small. To sum up, it is better to choose a group of images for which both the sum of consistency values and the sum of areas within the consistency square is small. Here, selecting a group of images with high consistency values within the consistency square is equivalent to selecting a group with low inconsistency. In summary, a group of reference images should be selected to maximize the total values of inconsistencies in consistency squares divided by the sum of the areas of consistency squares.</p><p>In practice, a procedure of reference image selection is the following. Suppose there are 𝑁 images in total, and we want to select one image as a reference image.</p><p>At first, baseline ranking is constructed as introduced in the previous subsection and obtains consistency values 𝐶𝑖,𝑗 (𝑖, 𝑗 &lt; 𝑁 ) for all pairs in the ranking. When evaluation space is divided into two with image 𝐼𝑚(𝑚 &lt; 𝑁 ), the sum of inconsistency values of the two spaces divided by the sum of the area is calculated as:</p><formula xml:id="formula_6">𝐷 (2) 𝑚 = ∑︀ 𝑚 𝑖=1 ∑︀ 𝑚 𝑗=1 (1 − 𝐶𝑖,𝑗) + ∑︀ 𝑁 𝑖=𝑚 ∑︀ 𝑁 𝑗=𝑚 (1 − 𝐶𝑖,𝑗) 𝑚 2 + (𝑁 − 𝑚) 2 .</formula><p>(7) By searching the position whose 𝐷 𝑚 are maximum, the best reference image 𝐼𝑚 that divides evaluation space into two can be obtained. Similar to this, by calculating the sum of inconsistency in the consistency squares divided by the sum of the area with different division numbers 𝑁 𝑟𝑒𝑓 + 1, 𝑁 𝑟𝑒𝑓 reference images can be obtained. When it comes to calculating these values, we apply a scheme of dynamic programming to reduce calculation costs.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Experiment</head><p>We conducted an experiment to evaluate the following things:</p><p>• How a proposed network can evaluate image pair.</p><p>• How appropriately reference images can be selected regarding consistency of both network and human annotators' evaluations. • How the selected reference images evaluate face images.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.">Dataset construction</head><p>At first, the face image dataset is constructed by capturing participants' face images. We conducted two types of experiments to construct datasets with different situations. In the first type of dataset, we asked a participant to sit in front of the camera and to listen to funny radio.</p><p>In the second type of dataset, we asked a participant to sit in front of the laptop PC and play a simple game. We captured facial images of these participants during the experiment. The second experiment was still experimental but closer to a natural scene than the first one. Each dataset was constructed only by one participant because our focus was to build a model to evaluate each individual. We collected two datasets of the first type, and one dataset of the second type. Then, we added labels between image pairs that showed which of the two images expressed more smiles. The annotation between images with a slight difference in the degree of smiling was difficult, even for humans. It might cause a mistake in giving the correct labels. Therefore, we utilized image pairs with a clear difference as training data in this paper. In particular, we manually annotated segments that we thought the degree of smiling ascended or descended monotonically. We then picked each segment's start and end frame to construct one pair with its label. The number of image pairs of each dataset were 216, 174, and 123, respectively. Also, all face images in these pairs were utilized as baseline images. That is, the size of the dataset is twice the size of the image pairs; 432, 348, 246, respectively.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.">Evaluation scheme</head><p>The procedure of evaluation was constructed the following four steps; (1) the proposed network was trained for each individual by collected data and was evaluated by the cross-validation scheme; (2) we constructed the baseline ranking and evaluated the voting-based algorithm by determining the rank within the images in the baseline ranking; (3) reference images were selected from baseline images as proposed in the previous section and evaluated by human annotators on how they were consistent; (4) we confirmed the smiling intensity of face images in the evaluation space constructed by reference images.</p><p>As for training our network, we utilized pre-trained feature extraction layers of VGG-Face <ref type="bibr" target="#b11">[12]</ref>, which was trained on millions of face images for person identification, and trained only fully connected layers.</p><p>Regarding the evaluation of the voting-based algorithm, the rank of each baseline image was determined by the baseline ranking itself. In this evaluation, the grand truth of the rank of each baseline image was given as the original rank of baseline ranking.</p><p>As for evaluating selected reference images, nine reference images were selected from the baseline images. Then, human annotators were asked to evaluate which of the two images represented more smiling for the pair of neighboring images in reference images. The images up to the third nearest neighbor images were considered a pair, and all image pairs were annotated twice by swapping the left and right sides of the comparison image. After comparing the image pairs within each dataset, the participants moved on to the next dataset. The order of evaluated image pairs was randomized for each dataset, but the order of the datasets was constant. We evaluated the consistency of the reference images by how accurate and consistent annotators evaluated the image pair. A group of 9 images regularly extracted every 𝑁/10 from the baseline images {𝐼 𝑟𝑎𝑛𝑘 1 , . . . , 𝐼 𝑟𝑎𝑛𝑘 9 } was used as the reference image for comparison. Seven participants between the ages of 21 and 27 (6 male and 1 female) were recruited as annotators.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.">Results</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.1.">Prediction accuracy</head><p>We first show the evaluation results of the trained comparison network in terms of accuracy.</p><p>In this evaluation, five-fold cross-validation was applied, and prediction accuracies of three datasets were 99.5% (215/216), 100% (174/174), 98.3% (121/123), respectively. Figure <ref type="figure" target="#fig_9">8</ref> shows examples of prediction results with Gradient-weighted Class Activation Mapping (Grad CAM) <ref type="bibr" target="#b12">[13]</ref>. In each figure, the first row shows the grand truth label and estimation results, and the second row shows regions on which a network focuses for the prediction as a heat map. From these results, we can see that the network returns accurate prediction results by correctly focusing on face regions, including the mouth and eyes, which are well known as corresponding to smiling, even from the small dataset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.2.">Consistency of voting based evaluation</head><p>Figure <ref type="figure" target="#fig_10">9</ref> shows four examples of estimated ranks of images in the baseline ranking, rank 1, 100, 200, and 400 of dataset 1. The total number of baseline images was 216 × 2 = 432. Each image was estimated as rank 1, 103, 199, and 398, respectively, and an almost correct rank can be estimated by the voting algorithm. Figure <ref type="figure" target="#fig_11">10</ref> shows all pairs of estimated rank and grand truth label of this evaluation in dataset 1. These results show the consistency and effectiveness of the voting-based algorithm, as it predicted almost consistent values for all baseline images.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.3.">Selected reference images</head><p>Figure <ref type="figure" target="#fig_5">11</ref> shows selected reference images by the proposed algorithm and equally picked up from the baseline ranking. The consistency table correlated to this result of dataset 1 is shown in Figure <ref type="figure" target="#fig_5">12</ref>. In this figure, the green line shows where the algorithm divides baseline ranking. These results show that there still appears to be some ambiguity between adjacent images, even with the proposed approach. However, it appears to be reduced compared to a group of images acquired at regular intervals.</p><p>The consistency table calculated by these selected reference images is shown in Figure <ref type="figure" target="#fig_13">13</ref>. We can see that almost all the cells represent blue. This result shows that ambiguities within reference images are small. Figure <ref type="figure" target="#fig_14">14</ref> and Figure <ref type="figure" target="#fig_5">15</ref> show the quantitative evaluation results of selected reference images by annotators. In each figure, "proposed" and "baseline" represent the result for the images up to the third nearest neighbor reference images of the proposed algorithm and baseline algorithm, respectively, and "proposed_adjacent" and "baseline_adjacent" represent the result for the images only the nearest neighbor reference images. That is, the difficulty of the evaluation becomes hard. Figure <ref type="figure" target="#fig_14">14</ref> shows a prediction accuracy of evaluation results. Here, we consider the order given by the proposed network as the grand truth of the prediction. Therefore, this result also shows a correlation between network prediction and human perceptions. Figure <ref type="figure" target="#fig_5">15</ref> shows the consistency of each participant's evaluation. In particular, it shows how much the same evaluation was given when the same image pair displayed with the left and right sides swapped. This high consistency indicates a low degree of ambiguity between image pairs. In almost all cases, reference images selected by the proposed algorithm obtain higher accuracies and higher consistencies. Since the smiles expressed in the experimental time were quantized into ten levels and the maximum value of the smile was not very high, both methods have a certain degree of similarity between the neighboring reference image pairs. Therefore, evaluation by humans may be somewhat difficult even with the proposed method. However, even in such a situation, we can confirm that the proposed method selects  image pairs with higher accuracy than the comparison method.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.4.">Smiling intensity evaluation</head><p>Examples of face images evaluated by selected reference images and the proposed network are shown in Figure <ref type="figure" target="#fig_4">16</ref>. Since it is sometimes hard to qualitatively evaluate two adjacent images in a row, the four reference images skip one rank at a time. The images with a smile level one class lower than the reference image are listed, and each row shows the same evaluation value. The images on the left side of the figure are recognized as having a higher degree of smiling. This result confirms that the proposed method effectively evaluates the degree of smiling within the ordinal scale. Finally, a part of the transition of the smiling intensity during the experiment is shown in Figure <ref type="figure" target="#fig_7">17</ref>. In this result, an evaluation score was smoothed by the median filter to trace the trend of transitions. We can see that the participant smiles several times in this period. It can be seen that smiles of slightly stronger intensity than the middle level occurred several times in succession in the first half of this period. In comparison, smiles of considerably stronger intensity occurred with a short interval in the second half of this period.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusion</head><p>In this paper, we propose an approach to evaluate the degree of smiling of individuals by ordinal scales based on multiple comparisons for the purpose of monitoring individuals. Suppose that we have enough data from individual face images; we also propose an algorithm for selecting appropriate reference images for the ordinal evaluation.</p><p>Experimental results show that our ordinal scale-based evaluation can successfully give the degree of not only clear smiling but also intermediate facial expressions. In addition, we can see that an evaluation space constructed by selected reference images by our algorithm is more consistent and, therefore, considered to be reasonable.</p><p>One of the future works is to map the proposed and constructed ordinal scale to some physical index. Although this paper proposed a method of selecting reference images that are somewhat reasonable when evaluated by humans, the validity of the scale would be improved if it could be mapped to some physical index. For example, by measuring the myoelectricity of facial muscles, the degree of muscle activity could be used as an index. In addition, the other future work is to apply this technique to people whose facial expressions do not change much, e.g., dementia patients, as we described in the introduction section.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Ethics</head><p>Our method aims to monitor the daily health conditions of a specific individual by evaluating the smiling intensity using a model trained specifically for the individual's facial images. Since data for model training and smiling intensity evaluation can be collected and processed at terminals installed in each individual's environment, it is expected to reduce the risk of leakage of particularly strong personal information such as facial images being stored in the cloud in practical applications.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 2 :Figure 3 :</head><label>23</label><figDesc>Figure 2: An example of sudden jump of evaluation scores for intermediate facial expressions by traditional facial expression recognition technique</figDesc><graphic coords="2,89.29,196.25,203.36,59.05" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Siamese-based network to compare face images to evaluate the degree of smiling</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 5 :</head><label>5</label><figDesc>Figure 5: Voting-based evaluation</figDesc><graphic coords="4,89.29,84.19,203.37,90.09" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>1 ,Figure 6 :</head><label>16</label><figDesc>Figure6(a) shows a comparison table of the result of all pair-wise comparisons in baseline images. Each color shows a probability of how a target image has a stronger smile than a reference image, i.e., 𝑃 𝐴𝑠 (𝑡𝑔𝑡, 𝑟𝑒𝑓 ). The blue shows a pair whose target image has a stronger smile than the reference image, i.e., 𝑃 𝐴𝑠 (𝑡𝑔𝑡, 𝑟𝑒𝑓 ) &gt; 𝑃 𝐷𝑒𝑠 (𝑡𝑔𝑡, 𝑟𝑒𝑓 ). In contrast, the red area shows a pair whose reference image has a stronger smile than a target image, i.e., 𝑃 𝐴𝑠 (𝑡𝑔𝑡, 𝑟𝑒𝑓 ) &lt; 𝑃 𝐷𝑒𝑠 (𝑡𝑔𝑡, 𝑟𝑒𝑓 ). The white area represents that the target and the reference images represent similar facial expressions.By sorting the baseline images based on the sum of the probability values in each column of this table, a baseline ranking considering the consistency of the strong-weak relationship can be constructed (Figure6 (b)). In particular, suppose we have 𝑁 baseline images {𝐼1, . . . , 𝐼𝑁 } in total, and denote images sorted in descending order by smiling intensity as {𝐼 𝑟𝑎𝑛𝑘 1 , . . . , 𝐼 𝑟𝑎𝑛𝑘 𝑁 }. Since the</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>1 ,</head><label>1</label><figDesc></figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head></head><label></label><figDesc>A consistency of strong-weak relations in the baseline ranking images (b) A consistency square of neighboring images</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_7"><head>Figure 7 :</head><label>7</label><figDesc>Figure 7: Consistencies of large small relationships in neighbor images as a part of consistency table in baseline ranking.</figDesc><graphic coords="5,431.83,101.96,61.36,61.36" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_9"><head>Figure 8 :</head><label>8</label><figDesc>Figure 8: Examples of prediction results by comparison network. The first row shows label and prediction results, and the second row shows regions on which a network focuses by Grad-CAM.</figDesc><graphic coords="6,408.39,183.22,93.31,92.75" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_10"><head>Figure 9 :</head><label>9</label><figDesc>Figure 9: Selected reference images of dataset 1. More smile images are located on the left side.</figDesc><graphic coords="7,302.62,84.19,203.36,72.28" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_11"><head>Figure 10 :</head><label>10</label><figDesc>Figure 10: Consistency of estimated rank and original rank in baseline ranking</figDesc><graphic coords="7,343.29,197.70,122.02,65.17" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_12"><head>Figure 11 :Figure 12 :</head><label>1112</label><figDesc>Figure 11: Selected reference images of dataset 1. More smile images are located on the left side.</figDesc><graphic coords="8,138.93,411.14,106.80,106.84" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_13"><head>Figure 13 :</head><label>13</label><figDesc>Figure 13: Consistency table calculated by selected reference images of dataset 1.</figDesc><graphic coords="8,333.12,84.19,142.36,147.62" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_14"><head>Figure 14 :</head><label>14</label><figDesc>Figure 14: Accuracies of annotators evaluation. Adjacent means the image pair consists of the nearest neighbor images.</figDesc><graphic coords="8,312.79,273.04,183.03,137.27" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_15"><head>Figure 15 :Figure 16 :</head><label>1516</label><figDesc>Figure 15: Consistencies of annotators evaluation of the same image pair. Adjacent means the image pair consists of the nearest neighbor images.</figDesc><graphic coords="9,99.46,84.19,183.03,137.27" type="bitmap" /></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Siamese-structure deep neural network recognizing changes in facial expression according to the degree of smiling</title>
		<author>
			<persName><forename type="first">K</forename><surname>Kondo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Nakamura</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Nakamura</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Satoh</surname></persName>
		</author>
		<idno type="DOI">10.1109/ICPR48806.2021.9411988</idno>
	</analytic>
	<monogr>
		<title level="m">Proc. of ICPR2020</title>
				<meeting>of ICPR2020</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="4605" to="4612" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<author>
			<persName><forename type="first">P</forename><surname>Ekman</surname></persName>
		</author>
		<ptr target="https://ci.nii.ac.jp/naid/10025007347/" />
		<title level="m">Facial action coding system (facs), A Human Face</title>
				<imprint>
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<author>
			<persName><forename type="first">B</forename><surname>Amos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Ludwiczuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Satyanarayanan</surname></persName>
		</author>
		<idno>CMU- CS-16-118</idno>
		<title level="m">Open-Face: A general-purpose face recognition library with mobile applications</title>
				<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
		<respStmt>
			<orgName>CMU School of Computer Science</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Technical Report</note>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Very deep convolutional networks for large-scale image recognition</title>
		<author>
			<persName><forename type="first">K</forename><surname>Simonyan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Zisserman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of ICLR 2015</title>
				<meeting>of ICLR 2015<address><addrLine>San Diego, CA, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2015">May 7-9, 2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Transfer learning technique with vgg-16 for near-infrared facial expression recognition</title>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">C</forename><surname>Atabansi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Cao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Xu</surname></persName>
		</author>
		<idno type="DOI">10.1088/1742-6596/1873/1/012033</idno>
		<idno>doi:</idno>
		<ptr target="10.1088/1742-6596/1873/1/012033" />
	</analytic>
	<monogr>
		<title level="j">Journal of Physics: Conference Series</title>
		<imprint>
			<biblScope unit="volume">1873</biblScope>
			<biblScope unit="page">12033</biblScope>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Facial expression recognition model based on improved vggnet</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<idno type="DOI">10.1109/ICECAI58670.2023.10177007</idno>
	</analytic>
	<monogr>
		<title level="m">2023 4th International Conference on Electronic Communication and Artificial Intelligence (ICECAI)</title>
				<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="404" to="408" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Signature verification using a &quot;siamese&quot; time delay neural network</title>
		<author>
			<persName><forename type="first">J</forename><surname>Bromley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Guyon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Lecun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Säckinger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Shah</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of NIPS&apos;93</title>
				<meeting>of NIPS&apos;93</meeting>
		<imprint>
			<date type="published" when="1993">1993</date>
			<biblScope unit="page" from="737" to="744" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Signature verification using a &quot;siamese&quot; time delay neural network</title>
		<author>
			<persName><forename type="first">J</forename><surname>Bromley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Bentz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Bottou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Guyon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Lecun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Moore</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Sackinger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Shah</surname></persName>
		</author>
		<idno type="DOI">10.1142/S0218001493000339</idno>
	</analytic>
	<monogr>
		<title level="j">International Journal of Pattern Recognition and Artificial Intelligence</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="page">25</biblScope>
			<date type="published" when="1993">1993</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Siamese neural network based few-shot learning for anomaly detection in industrial cyber-physical systems</title>
		<author>
			<persName><forename type="first">X</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Liang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Shimizu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Jin</surname></persName>
		</author>
		<idno type="DOI">10.1109/TII.2020.3047675</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Industrial Informatics</title>
		<imprint>
			<biblScope unit="volume">17</biblScope>
			<biblScope unit="page" from="5790" to="5798" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Siamese neural networks for one-shot image recognition</title>
		<author>
			<persName><forename type="first">G</forename><surname>Koch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Zemel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Salakhutdinov</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the deep learning workshop in the 32nd International Conference on Machine Learning</title>
				<meeting>of the deep learning workshop in the 32nd International Conference on Machine Learning</meeting>
		<imprint>
			<date type="published" when="2015">2015</date>
			<biblScope unit="volume">2</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Facial expression change recognition on neutralnegative axis based on siamese-structure deep neural network</title>
		<author>
			<persName><forename type="first">J</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Shimonishi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Kondo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Nakamura</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Cross-Cultural Design. Product and Service Design, Mobility and Automotive Design, Cities, Urban Areas, and Intelligent Environments Design: 14th International Conference, CCD 2022, Held as Part of the 24th HCI International Conference</title>
				<meeting><address><addrLine>HCII</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">2022. 2022</date>
			<biblScope unit="page" from="583" to="598" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Deep face recognition</title>
		<author>
			<persName><forename type="first">O</forename><forename type="middle">M</forename><surname>Parkhi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Vedaldi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Zisserman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">British Machine Vision Conference</title>
				<imprint>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Grad-cam: Visual explanations from deep networks via gradient-based localization</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">R</forename><surname>Selvaraju</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Cogswell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Das</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Vedantam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Parikh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Batra</surname></persName>
		</author>
		<idno type="DOI">10.1007/s11263-019-01228-7</idno>
	</analytic>
	<monogr>
		<title level="j">International Journal of Computer Vision</title>
		<imprint>
			<biblScope unit="volume">128</biblScope>
			<biblScope unit="page" from="336" to="359" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
