<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Automated Pain Detection in Facial Videos of Children using Human-Assisted Transfer Learning</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Xiaojing</forename><surname>Xu</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Department of Electrical and Computer Engineering</orgName>
								<orgName type="institution">University of California San Diego</orgName>
								<address>
									<addrLine>La Jolla</addrLine>
									<region>CA</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Kenneth</forename><forename type="middle">D</forename><surname>Craig</surname></persName>
							<email>kcraig@psych.ubc.ca</email>
							<affiliation key="aff1">
								<orgName type="department">Department of Psychology</orgName>
								<orgName type="institution">University of British Columbia Vancouver</orgName>
								<address>
									<region>BC</region>
									<country key="CA">Canada</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Damaris</forename><surname>Diaz</surname></persName>
							<affiliation key="aff2">
								<orgName type="department">Rady Childrens Hospital and Department of Pediatrics</orgName>
								<orgName type="institution">University of California San Diego</orgName>
								<address>
									<region>CA</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Matthew</forename><forename type="middle">S</forename><surname>Goodwin</surname></persName>
							<email>m.goodwin@northeastern.edu</email>
							<affiliation key="aff3">
								<orgName type="department">Department of Health Sciences</orgName>
								<orgName type="institution">Northeastern University</orgName>
								<address>
									<settlement>Boston</settlement>
									<region>MA</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Murat</forename><surname>Akcakaya</surname></persName>
							<email>akcakaya@pitt.edu</email>
							<affiliation key="aff4">
								<orgName type="department">Department of Electrical and Computer Engineering</orgName>
								<orgName type="institution">University of Pittsburgh</orgName>
								<address>
									<settlement>Pittsburgh</settlement>
									<region>PA</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Büşra</forename><forename type="middle">Tugçe</forename><surname>Susam</surname></persName>
							<affiliation key="aff4">
								<orgName type="department">Department of Electrical and Computer Engineering</orgName>
								<orgName type="institution">University of Pittsburgh</orgName>
								<address>
									<settlement>Pittsburgh</settlement>
									<region>PA</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Jeannie</forename><forename type="middle">S</forename><surname>Huang</surname></persName>
							<email>jshuang@ucsd.edu</email>
							<affiliation key="aff2">
								<orgName type="department">Rady Childrens Hospital and Department of Pediatrics</orgName>
								<orgName type="institution">University of California San Diego</orgName>
								<address>
									<region>CA</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Virginia</forename><forename type="middle">R</forename><surname>De Sa</surname></persName>
							<email>desa@ucsd.edu</email>
							<affiliation key="aff5">
								<orgName type="department">Department of Cognitive Science</orgName>
								<orgName type="institution">University of California San Diego</orgName>
								<address>
									<addrLine>La Jolla</addrLine>
									<region>CA</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Automated Pain Detection in Facial Videos of Children using Human-Assisted Transfer Learning</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">A1DB2766E85622313C294B884AFAE10A</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T22:31+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>automated pain detection</term>
					<term>transfer learning</term>
					<term>facial action units</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Accurately determining pain levels in children is difficult, even for trained professionals and parents. Facial activity provides sensitive and specific information about pain, and computer vision algorithms have been developed to automatically detect Facial Action Units (AUs) defined by the Facial Action Coding System (FACS). Our prior work utilized information from computer vision, i.e. automatically detected facial AUs to develop classifiers to distinguish between pain and no-pain conditions. However, application of pain/no-pain classifiers based on automated AU codings across different environmental domains resulted in diminished performance. In contrast, classifiers based on manually coded AUs demonstrated reduced environmentally-based variability in performance. To improve classification performance in the current work, we applied transfer learning by training another machine learning model to map automated AU codings to a subspace of manual AU codings to enable more robust pain recognition performance when only automatically coded AUs are available for the test data. With this transfer learning method, we improved the Area under the ROC Curve (AUC) on independent data (new participants) from our target data domain from 0.69 to 0.72.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>In the classic model of machine learning, scientists train models on training data to accurately detect a desired outcome and apply learned models to new data measured under identical circumstances to validate their performance. Given the real world and its notable variation, it is tempting to apply learned models to data measured under similar but not identical circumstances. However, performance in such circumstances often deteriorates because of unmeasured factors not accounted for between original and new datasets. Nevertheless, lessons can be learned from similar scenarios. Transfer learning or inductive transfer in machine learning focuses on storing knowledge gained while solving one problem and applying it to a different but related problem <ref type="bibr" target="#b0">[1]</ref>. We describe application of transfer learning to the important clinical problem of pain detection in children.</p><p>Accurate measurement of pain severity in children is difficult, even for trained professionals and parents. This is a critical problem as over-medication can result in adverse side-effects, including opioid addiction, and under-medication can lead to unnecessary suffering <ref type="bibr" target="#b1">[2]</ref>.</p><p>The current clinical gold standard and most widely employed method of assessing clinical pain is patient self-report <ref type="bibr" target="#b2">[3]</ref>. However, this method is subjective and vulnerable to bias. Consequently, clinicians often distrust pain self-reports, and find them more useful for comparisons over time within individuals, rather than comparisons between individuals <ref type="bibr" target="#b3">[4]</ref>. Further, infants and older children with communication/neurologic disabilities do not have the ability or capacity to self-report pain levels <ref type="bibr" target="#b2">[3]</ref>[5] <ref type="bibr" target="#b5">[6]</ref>. As a result, to evaluate pain in populations with communication limitations, observational tools based on behavioral nonverbal indicators associated with pain have been developed <ref type="bibr" target="#b6">[7]</ref>.</p><p>Of the various modalities of nonverbal expression (e.g., bodily movement, vocalizations), facial activity can provide the most sensitive, specific, and accessible information about the presence, nature, and severity of pain across the life span, from infancy <ref type="bibr" target="#b7">[8]</ref> through to advanced age <ref type="bibr" target="#b8">[9]</ref>. Observers largely consider facial activity during painful events to be a relatively spontaneous reaction <ref type="bibr" target="#b6">[7]</ref>.</p><p>Evaluation of pain based on facial indicators requires two steps: (1) Extraction of facial pain features and (2) pain recognition based on these features. For step <ref type="bibr" target="#b0">(1)</ref>, researchers have searched for reliable facial indicators of pain, such as the anatomically-based, objectively coded Facial Action Units (AUs) defined by the Facial Action Coding System (FACS) <ref type="bibr" target="#b9">[10]</ref> <ref type="bibr" target="#b10">[11]</ref> (Visualizations of the facial activation units can be found at https://imotions.com/blog/facial-action-codingsystem/). However, identifying these AUs traditionally requires time intensive offline coding by trained human coders, limiting their application in real-time clinical settings. Recently, algorithms to automatically detect AUs <ref type="bibr" target="#b10">[11]</ref> have been developed and implemented in software such as iMotions (imotions.com) allowing automatical output of AU probabilities in real-time based on direct recording of face video. In step (2), Linear models <ref type="bibr" target="#b4">[5]</ref>, SVM <ref type="bibr" target="#b11">[12]</ref> and Neural Networks <ref type="bibr" target="#b12">[13]</ref> have been used to recognize pain based on facial features. In this paper, we first combine iMotions and Neural Networks to build an automated pain recognition model.</p><p>Although a simple machine learning model based on features extracted by a well-designed algorithm can perform well when training data and test data have similar statistical properties, problems arise when the data follow different distributions. We discovered this issue when training videos were recorded in one environment/setting and test videos in another. One way to deal with this problem is to use transfer learning, which discovers "common knowledge" across domains and uses this knowledge to complete tasks in a new domain with a model learned in the old domain <ref type="bibr" target="#b13">[14]</ref>. In this paper, we show that features extracted from human-coded (manual) AU codings are less sensitive to domain changes than features extracted from iMotions (automated) AU codings, and thus develop a simple method that learns a projection from automated features onto a subspace of manual features. Once this mapping is learned, future automatically coded data can be automatically transformed to a representation that is more robust between domains.</p><p>To summarize, our contributions in this paper include:</p><p>-Demonstration that environmental factors modulate the ability of automatically coded AUs to recognize clinical pain in videos -Demonstration that manually coded AUs (especially previously established "pain-related" ones) can be used to successfully recognize pain in video with machine learning across different environmental domains -Development of a transfer learning method to transfer automated features to the manual feature space that improves automatic recognition of clinical pain across different environmental domains 2 Methods</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Participants</head><p>143 pediatric research participants (94 males, 49 females) aged 12 <ref type="bibr" target="#b9">[10,</ref><ref type="bibr" target="#b14">15]</ref> (median [25%, 75%]) years old and primarily Hispanic (78%) who had undergone medically necessary laparoscopic appendectomy were videotaped for facial expressions during surgical recovery. Participating children had been hospitalized following surgery for post-surgical recovery and were recruited for participation within 24 hours of surgery at a pediatric tertiary care center. Exclusion criteria included regular opioid use within the past 6 months, documented mental or neurologic deficits preventing study protocol compliance, and any facial anomaly that might alter computer vision facial expression analysis. Parents provided written informed consent and youth gave written assent. The local institutional review board approved the research protocol.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Experimental Design and Data Collection</head><p>Data were collected over 3 visits (V): V1 within 24 hours after appendectomy; V2 within the calendar day after the first visit; and V3 at a follow-up visit 25 <ref type="bibr">[19,</ref><ref type="bibr">28]</ref> (median [25%, 75%]) days postoperatively when pain was expected to have fully subsided. Data were collected in two environmental conditions: V1 and V2 in hospital and V3 in the outpatient lab. At every visit, two 10-second videos (60 fps at 853x480 pixel resolution) of the face were recorded while manual pressure was exerted at the surgical site for 10 seconds (equivalent of a clinical examination). In the hospital visits (V1, V2), the participants were lying in the hospital bed with the head of the bed raised. In V3, they were seated in a reclined chair. Participants rated their pain level during the pressure using a 0-10 Numerical Rating Scale, where 0 = no-pain and 10 = worst pain ever. Following convention for clinically significant pain <ref type="bibr" target="#b14">[15]</ref>, videos with pain ratings of 0-3 were labeled as no-pain, and videos with pain ratings of 4-10 were labeled as pain, for classification purposes. 324 pain videos were collected from V1/2, 195 no-pain videos were collected from V1/2, and 235 no-pain videos were collected from V3. Figure <ref type="figure" target="#fig_0">1</ref> demonstrates the distribution of pain and no-pain videos across environmental domains.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3">Feature Extraction</head><p>For each 10-second video sample, we extracted AU codings per frame to obtain a sequence of AUs. This was done both automatically by iMotions software (www.imotions.com) and manually by a trained human in a limited subset. We then extracted features from the sequence of AUs.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Automated Facial Action Unit Detection:</head><p>The iMotions software integrates Emotient's FACET technology (www.imotions.com/emotient) which was formally known as CERT <ref type="bibr" target="#b15">[16]</ref>. In the described work, the iMotions software was used to process the videos to automatically extract 20 AUs (AU 1, 2, 4, 5, 6, 7, 9, 10, 12, 14, 15, 17, 18, 20, 23, 24, 25, 26, 28, 43) and three head pose indicators (yaw, pitch and roll) from each frame. The values of these codings are the estimated log probabilities of AUs, ranging from -4 to 4.</p><p>Manual Facial Action Unit Detection: A trained human FACS AU coder manually coded 64 AUs (AU1-64) for each frame of a subset of videos by labeling the AU intensities (0-5, 0 = absence).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Feature Dimension Reduction:</head><p>The number of frames in our videos were too large to use full sequences of frame-coded AUs. To reduce dimensionality, we applied 11 statistics (mean, max, min, standard deviation, 95th, 85th, 75th, 50th, 25th percentiles, half-rectified mean, and max-min) to each AU over all frames to obtain 11 × 23 features for automatically coded AUs, and 11 × 64 features for manually coded AUs. We call these automated features and manual features, respectively. The range of each feature was rescaled to [0, 1] to normalize features over the training data. Neural Network Model to Predict Manual Features with Automated Features: A neural network with the same structure was used to predict manual features from automated features, except that the output layer was linear and mean squared error was used as the loss function.</p><p>Model Training and Testing: Experiments were conducted in a participantbased (each participant restricted to one fold) 10-fold cross-validation fashion.</p><p>Participants were divided into 10 folds, and each time 1 fold was used as the test set, and the other 9 folds together were used as the training set. We balanced classes for each participant in each training set by duplicating samples from the under-represented class. 1/9 participants in the training set were picked randomly as a nested-validation set for early stopping in the neural network training. A batch size of 1/8 the size of training set was used. We examined the receiver operating characteristic curve (ROC curve) which plots True Positive Rate against False Positive Rate as the discrimination threshold is varied. We used the Area under the Curve (AUC) to evaluate the performance of classifiers. We considered data from 3 domains (D) as shown in Figure <ref type="figure" target="#fig_0">1</ref>: (1) D1 with pain and no-pain both from V1/2 in hospital, (2) D2 with pain from V1/2 in hospital and no-pain from V3 from outpatient lab, and (3) All data, i.e., pain from V1/2 and no-pain from V1/2/3. The clinical goal was to be able to discriminate pain levels in the hospital; thus evaluation on D1 (where all samples were from the hospital bed) was the most clinically relevant evaluation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Analysis and Discussion</head><p>Data of 73 participants labeled by both human and iMotions were used through section 3.1 to 3.5, and data of the remaining 70 participants having only automated (iMotions) AU codings were used for independent test set evaluation in the results section. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Automated Classifier Performance Varies by Environment</head><p>Using automated features, we first combined all visit data and trained a classifier to distinguish pain from no-pain. This classifier performed well in general (AUC = 0.77 ± 0.011 on All data), but when we looked at different domains, the performance of D1 (the most clinically relevant in-Hospital environment) was much inferior to that on D2, as shown in data rows 1 and 4 under the "Automated" column in Table <ref type="table" target="#tab_0">1</ref>. There were two main differences between D1 and D2, i.e. between V1/2 and V3 no-pain samples. The first was that in V1/2, patients usually still had some pain and their self-ratings were greater than 0, while in V3, no-pain ratings were usually 0 reflecting a "purer" no-pain signal. The second difference was that V1/2 happened in hospital with patients in beds and V3 videos were recorded in an outpatient lab with the patient sitting in a reclined chair. Since automated recognition of AUs is known to be sensitive to facial pose and lighting differences, we hypothesized that the added discrepancy in classification performance between D1 and D2 was mainly due to the model classifying based on environmental differences between V1/2 and V3. In other words the classifier when trained and tested on D2, might be classifying "lying in hospital bed" vs "more upright in outpatient chair" as much as pain vs no-pain. (This is similar to doing well at recognizing cows by recognizing a green background).</p><p>In order to investigate this hypothesis and attempt to improve classification on the clinically relevant D1, we trained a classifier using only videos from D1. Within the "Automated" column, row 2 in Table <ref type="table" target="#tab_0">1</ref> shows that performance on automated D1 classification doesn't drop much when D2 samples are removed from the training set. At the same time, training using only D2 data results in the worst classification on D1 (row 3), but the best classification on D2 (last row) as the network is able to exploit the environmental differences (no-pain+more upright from V3, pain+lying-down from V1/2).</p><p>Figure <ref type="figure">2</ref> (LEFT) shows ROC curves of within and across domain tests for models trained on automated features in D2. The red dotted curve corresponds to testing on D2 (within domain) and the blue solid curve corresponds to testing on D1 (across domain). The model did well on within domain classification, but failed on across domain tasks.  Fig. <ref type="figure">2</ref>. ROC Curves for classification within and across domains using automated features (left), manual features (middle) and pain-related manual features (right). The red dotted lines are ROCs when the machine is able to use environment information to differentiate pain and no-pain conditions, and the blue solid lines show the machine's ability to discriminate between pain and no-pain based on AU information alone. The straight yellow line graphs the performance of random chance.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Classification Based on Manual AUs Are Less Sensitive to Environmental Changes</head><p>We also trained a classifier on manual AUs labeled by a human coder. Interestingly, results from the classifier trained on manual AUs showed less of a difference in AUCs between the domains, with a higher AUC for D1 and a lower AUC for D2 relative to those with the automated AUs (see Table <ref type="table" target="#tab_0">1</ref> "Manual" and "Automated" columns). The manual AUs appeared to be less sensitive to changes in the environment reflecting the ability of the human labeler to consistently code AUs without being affected by lighting and pose variations.</p><p>When we restricted training data from All to only D1 or only D2 data, classification performance using manual AUs went down, likely due to the reduction in training data, and training with D2 always gave better performance than training with D1 on both D1 and D2 test data, which should be the case since D2 is higher in "pain" quality. These results appear consistent with our hypothesis that human coding of AUs is not as sensitive as machine coding of AUs to the environmental differences between V1/2 and V3.</p><p>Figure <ref type="figure">2</ref> (MIDDLE) displays the ROC curves for manual features. As discussed above, in contrast to the plot on the left for automated features, manual coding performance outperformed automated coding performance in the clinically relevant test in D1. The dotted red curve representing within domain performance is only slightly higher than the solid blue curve, likely due in part to the quality difference in no-pain samples in V1/2 and V3 and also possibly due to any small amount of environmental information that the human labeler was affected by. Note that ignoring the correlated environmental information in D2 (pain faces were more reclined and no-pain faces were more upright) resulted in a lower numerical performance on D2 but does not likely reflect worse classification of pain. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Restricting Manual AUs to Those Associated with Pain Improves Classification</head><p>In an attempt to reduce the influence of environmental conditions to further improve accuracy on D1, we restricted the classifier to the eight AUs that have been consistently associated with pain: 4 (Brow Lowerer), 6 (Cheek Raiser), 7 (Lid Tightener), 9 (Nose Wrinkler), 10 (Upper Lip Raiser), 12 (Lip Corner Puller), 20 (Lip Stretcher), and 43 (Eyes Closed) <ref type="bibr" target="#b16">[17,</ref><ref type="bibr" target="#b17">18]</ref> to obtain 11 (statistics) ×8 (AUs) features. Pain prediction results using these "pain" features are shown in the last two columns in Table <ref type="table" target="#tab_0">1</ref>. Results show that using only pain-related AUs improved the classification performances of manual features. However, it did not seem to help as much for automated features. Similarly, Figure <ref type="figure">2</ref> (RIGHT) shows that limiting manual features to use only pain-related AUs further improved D1 performance when training with D2. We also performed PCA on these pain-related features and found that performance in the hospital environmental domain was similar if using 4 or more principal components.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4">iMotions AUs Are Different Than Manual FACS AUs</head><p>Computer Vision AU automatic detection algorithms have been programmed/ trained on manual FACS data. However, we demonstrated differential performance of AUs encoded automatically versus manually. To understand the relationship between automatically encoded v. manually coded AUs, we computed correlations between automatically coded AUs and manually coded AUs at the frame level as depicted in Figure <ref type="figure" target="#fig_2">3</ref>. If two sets of AUs were identical, the diagonal of the matrix (marked with small centered dots) should yield the highest correlations, which was not the case. For example, manual AU 6 was highly correlated with automated AU 12 and 14, but had relatively low correlation with automated AU 6.</p><p>The correlation matrix shows that not only are human coders less affected by environmental changes, the AUs they code are not in agreement with the automated AUs. (We separately had another trained human coder code a subset of the videos and observed closer correlation between the humans than between each human and iMotions). This likely explains the reduced improvement from restricting the automated features model to "pain-related AUs" as these have been determined based on human FACS coded AUs.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.5">Transfer Learning via Mapping to Manual Features Improves Performance</head><p>We have shown that manual codings are not as sensitive to domain change. However, manual coding of AUs is very time-consuming and not amenable to an automated real-time system. In an attempt to leverage manual coding to achieve similar robustness with automatic AUs, we utilized transfer learning and mapped automated features to the space of manual features. Specifically, we trained a machine learning model to estimate manual features from automated features using data coded by both iMotions and a human. Separate models were trained to predict: manual features of 64 AUs, manual features of the eight pain-related AUs, principal components (PCs) of the manual features of the eight pain-related AUs. PCA dimensionality reduction was used due to insufficient data for learning a mapping from all automated AUs to all manual AUs. Once the mapping network was trained, we used it to transform the automated features and train a new network on these transformed data for the pain/no-pain classification. The 10-fold cross-validation was done consistently so that the same training data was used to train the mapping network and the pain-classification network.</p><p>In Table <ref type="table" target="#tab_2">2</ref>, we show the classification AUCs when the classification model was trained and tested with outputs from the prediction network. We observed that when using All data to train (which had the best performance), with the transfer learning prediction network, automated features performed much better in classification on D1 (0.68−0.69 compared to 0.61−0.63 in Table <ref type="table" target="#tab_0">1</ref>). Predicting 4 principal components of manual pain-related features gave the best performance on our data. Overall, the prediction network helped in domain adaptation of a pain recognition model using automatically extracted AUs.</p><p>Figure <ref type="figure" target="#fig_4">5</ref> (LEFT) plots the ROC curves using the transfer learning classifier within and across domains trained and tested using 4 predicted features. Com- pared to Figure <ref type="figure">2</ref> (LEFT), the transferred automated features showed properties more like manual features, with smaller differences between performances on the two domains and higher AUC on the clinically relevant D1. Table <ref type="table" target="#tab_2">2</ref> shows numerically how transfer learning helped automated features to ignore environmental information in D2 like humans, and learn pure pain information which can also be used in classification on D1. Within domain classification performance for D1 was also improved with the prediction network. These results show that by mapping to the manual feature space, automated features can be promoted to perform better in pain classification.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Results</head><p>In the previous section, we showed that in Figure <ref type="figure" target="#fig_3">4</ref>, classification with painrelated pain features (2) performed better than automated features (1) on D1, which was the clinically relevant classification. We have also found that applying PCA to manual features (3-4) didn't change the performance on D1 much. Thus we introduced a transfer learning model to map automated features first to manual pain-related features (or the top few principal components of them), and then used the transferred features for classification (6-2, or 5-4). We got similar results to manual features on D1 with transfer learning model (5-4) mapping to 4 principal components of manual features.</p><p>In this section we report on the results from testing our transfer learning method on a new separate dataset (new participants), which has only automated features. Table <ref type="table" target="#tab_0">1</ref> shows that without our method, training on all data and restricting to pain-related AUs resulted in the best performance for D1. And cross-validation results in Table <ref type="table" target="#tab_2">2</ref> shows that with our method, predicting 4 PCs yielded the best performance for D1. With these optimal choices of model structure and training domain, we trained two models </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Conclusion</head><p>In the described work, we recognized differences in classifier model performance (on pain vs no-pain) across data domains that reflected environmental differences as well as differences reflecting how the data were encoded (automatically v. manually). We then introduced a transfer learning model to map automated features first to manual pain-related features (or principal components of them), and then used the transferred features for classification (6-2, or 5-4 in Figure <ref type="figure" target="#fig_3">4</ref>). This allowed us to leverage data from another domain to improve classifier performance on the clinically relevant task of distinguishing pain levels in the hospital.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Data Domain Illustration</figDesc><graphic coords="5,181.68,115.84,252.00,53.29" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head></head><label></label><figDesc>0</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>FrameFig. 3 .</head><label>3</label><figDesc>Fig. 3. Correlation matrix of AU pairs from automated and manual codings using All data</figDesc><graphic coords="8,258.62,124.51,90.58,84.42" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Fig. 4 .</head><label>4</label><figDesc>Fig.<ref type="bibr" target="#b3">4</ref>. Illustration of Machine Learning Models. 1/2 are classifications using automated/manual pain features, in which 2 does better than 1. 3-4 can be done to reduce feature dimensions while maintaining the performance. 6-2 and 5-4 are our transfer learning models, training a regression network to map automated features to a subspace of manual pain features before classification.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Fig. 5 .</head><label>5</label><figDesc>Fig. 5. ROC Curves for classification within and across domains using our transfer learning model (left) and plot of average model output pain score (with error bars indicating standard deviation) over true pain level (right)</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 .</head><label>1</label><figDesc>AUC for Classification with SEM (Standard Error of The Mean)</figDesc><table><row><cell cols="3">Train on Test on Automated</cell><cell>Manual</cell><cell>Automated "Pain" Features</cell><cell>Manual "Pain" Features</cell></row><row><cell>All</cell><cell cols="4">D1 0.61 ± 0.007 0.66 ± 0.007 0.63 ± 0.007</cell><cell>0.69 ± 0.006</cell></row><row><cell>D1</cell><cell>D1</cell><cell cols="3">0.6 ± 0.009 0.61 ± 0.008 0.62 ± 0.007</cell><cell>0.65 ± 0.007</cell></row><row><cell>D2</cell><cell cols="4">D1 0.58 ± 0.006 0.66 ± 0.006 0.61 ± 0.005</cell><cell>0.7 ± 0.005</cell></row><row><cell>All</cell><cell>D2</cell><cell cols="2">0.9 ± 0.005 0.8 ± 0.006</cell><cell>0.89 ± 0.005</cell><cell>0.8 ± 0.004</cell></row><row><cell>D1</cell><cell cols="4">D2 0.72 ± 0.006 0.69 ± 0.008 0.75 ± 0.008</cell><cell>0.74 ± 0.007</cell></row><row><cell>D2</cell><cell cols="4">D2 0.93 ± 0.005 0.79 ± 0.008 0.91 ± 0.003</cell><cell>0.8 ± 0.006</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 2 .</head><label>2</label><figDesc>AUC (and SEM) with Transferred Automated Features</figDesc><table><row><cell cols="2">Train on Test on All Features "Pain" Features</cell><cell>7 PCs</cell><cell>4 PCs</cell><cell>1 PC</cell></row><row><cell>All</cell><cell>D1 0.62 ± 0.007 0.64 ± 0.01</cell><cell cols="3">0.68 ± 0.009 0.69 ± 0.006 0.63 ± 0.008</cell></row><row><cell>D1</cell><cell cols="4">D1 0.64 ± 0.008 0.65 ± 0.009 0.67 ± 0.011 0.67 ± 0.01 0.63 ± 0.013</cell></row><row><cell>D2</cell><cell>D1 0.58 ± 0.006 0.59 ± 0.01</cell><cell cols="3">0.66 ± 0.007 0.68 ± 0.006 0.65 ± 0.006</cell></row><row><cell>All</cell><cell>D2 0.84 ± 0.011 0.83 ± 0.009</cell><cell cols="3">0.76 ± 0.01 0.75 ± 0.011 0.67 ± 0.012</cell></row><row><cell>D1</cell><cell cols="4">D2 0.69 ± 0.008 0.72 ± 0.013 0.71 ± 0.011 0.71 ± 0.013 0.66 ± 0.017</cell></row><row><cell>D2</cell><cell cols="4">D2 0.89 ± 0.004 0.85 ± 0.008 0.77 ± 0.007 0.74 ± 0.01 0.68 ± 0.009</cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This work was supported by National Institutes of Health National Institute of Nursing Research grant R01 NR013500 and by IBM Research AI through the AI Horizons Network. Many thanks to Ryley Unrau for manual FACS coding and Karan Sikka for sharing his code and ideas used in <ref type="bibr" target="#b4">[5]</ref>.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Spring research presentation: A theoretical foundation for inductive transfer</title>
		<author>
			<persName><forename type="first">Jeremy</forename><surname>West</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Dan</forename><surname>Ventura</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Sean</forename><surname>Warnick</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">College of Physical and Mathematical Sciences</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<date type="published" when="2007">2007</date>
		</imprint>
		<respStmt>
			<orgName>Brigham Young University</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Pain assessment in children with special needs: A review of the literature</title>
		<author>
			<persName><forename type="first">Esther</forename><surname>Brenna L Quinn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Laura</forename><surname>Seibold</surname></persName>
		</author>
		<author>
			<persName><surname>Hayman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Exceptional Children</title>
		<imprint>
			<biblScope unit="volume">82</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="44" to="57" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<author>
			<persName><forename type="first">Ghada</forename><surname>Zamzmi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Chih-Yun</forename><surname>Pai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Dmitry</forename><surname>Goldgof</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Rangachar</forename><surname>Kasturi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yu</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Terri</forename><surname>Ashmeade</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1607.00331</idno>
		<title level="m">Machine-based multimodal pain assessment tool for infants: a review</title>
				<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
	<note type="report_type">preprint</note>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Childrens self-report of pain intensity: what we know, where we are headed</title>
		<author>
			<persName><forename type="first">Von</forename><surname>Carl</surname></persName>
		</author>
		<author>
			<persName><surname>Baeyer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Pain Research and Management</title>
		<imprint>
			<biblScope unit="volume">14</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="39" to="45" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Automated assessment of childrens postoperative pain using computer vision</title>
		<author>
			<persName><forename type="first">Karan</forename><surname>Sikka</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Alex</forename><forename type="middle">A</forename><surname>Ahmed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Damaris</forename><surname>Diaz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Matthew</forename><forename type="middle">S</forename><surname>Goodwin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kenneth</forename><forename type="middle">D</forename><surname>Craig</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Marian</forename><forename type="middle">S</forename><surname>Bartlett</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jeannie</forename><forename type="middle">S</forename><surname>Huang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Pediatrics</title>
		<imprint>
			<biblScope unit="volume">136</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="e124" to="e131" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">The automatic detection of chronic pain-related expression: requirements, challenges and the multimodal emopain dataset</title>
		<author>
			<persName><forename type="first">Sebastian</forename><surname>Min Sh Aung</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Bernardino</forename><surname>Kaltwang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Brais</forename><surname>Romera-Paredes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Aneesha</forename><surname>Martinez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Matteo</forename><surname>Singh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Michel</forename><surname>Cella</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hongying</forename><surname>Valstar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Andrew</forename><surname>Meng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Moshen</forename><surname>Kemp</surname></persName>
		</author>
		<author>
			<persName><surname>Shafizadeh</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE transactions on affective computing</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="435" to="451" />
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Childrens behavioral pain cues: Implicit automaticity and control dimensions in observational measures</title>
		<author>
			<persName><forename type="first">Samantha</forename><forename type="middle">R</forename><surname>Kamal Kaur Sekhon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Judith</forename><surname>Fashler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Spencer</forename><surname>Versloot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kenneth</forename><forename type="middle">D</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><surname>Craig</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Pain Res Manag</title>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Pain expression in neonates: facial action and cry</title>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">E</forename><surname>Ruth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kenneth</forename><forename type="middle">D</forename><surname>Grunau</surname></persName>
		</author>
		<author>
			<persName><surname>Craig</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Pain</title>
		<imprint>
			<biblScope unit="volume">28</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="395" to="410" />
			<date type="published" when="1987">1987</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Pain assessment in elderly adults with dementia</title>
		<author>
			<persName><forename type="first">Thomas</forename><surname>Hadjistavropoulos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Keela</forename><surname>Herr</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kenneth</forename><forename type="middle">D</forename><surname>Kenneth M Prkachin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Stephen</forename><forename type="middle">J</forename><surname>Craig</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Albert</forename><surname>Gibson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jonathan</forename><forename type="middle">H</forename><surname>Lukas</surname></persName>
		</author>
		<author>
			<persName><surname>Smith</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">The Lancet Neurology</title>
		<imprint>
			<biblScope unit="volume">13</biblScope>
			<biblScope unit="issue">12</biblScope>
			<biblScope unit="page" from="1216" to="1227" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Measuring facial movement</title>
		<author>
			<persName><forename type="first">Paul</forename><surname>Ekman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Wallace</forename><forename type="middle">V</forename><surname>Friesen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Environmental psychology and nonverbal behavior</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="56" to="75" />
			<date type="published" when="1976">1976</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Automatic analysis of facial actions: A survey</title>
		<author>
			<persName><forename type="first">Brais</forename><surname>Martinez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Michel</forename><forename type="middle">F</forename><surname>Valstar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Bihan</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Maja</forename><surname>Pantic</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Trans on Affective Computing</title>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">The painful face-pain expression recognition using active appearance models</title>
		<author>
			<persName><forename type="first">Ahmed</forename><surname>Bilal Ashraf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Simon</forename><surname>Lucey</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jeffrey</forename><forename type="middle">F</forename><surname>Cohn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Tsuhan</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Zara</forename><surname>Ambadar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Patricia</forename><forename type="middle">E</forename><surname>Kenneth M Prkachin</surname></persName>
		</author>
		<author>
			<persName><surname>Solomon</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Image and vision computing</title>
		<imprint>
			<biblScope unit="volume">27</biblScope>
			<biblScope unit="issue">12</biblScope>
			<biblScope unit="page" from="1788" to="1796" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Pain recognition using artificial neural network</title>
		<author>
			<persName><forename type="first">Md</forename><surname>Maruf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Monwar</forename></persName>
		</author>
		<author>
			<persName><forename type="first">Siamak</forename><surname>Rezaei</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Signal Processing and Information Technology</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2006">2006. 2006</date>
			<biblScope unit="page" from="28" to="33" />
		</imprint>
	</monogr>
	<note>IEEE International Symposium on</note>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">A survey on transfer learning</title>
		<author>
			<persName><forename type="first">J</forename><surname>Sinno</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Qiang</forename><surname>Pan</surname></persName>
		</author>
		<author>
			<persName><surname>Yang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Trans on knowledge and data engineering</title>
		<imprint>
			<biblScope unit="volume">22</biblScope>
			<biblScope unit="issue">10</biblScope>
			<biblScope unit="page" from="1345" to="1359" />
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">How do changes in pain severity levels correspond to changes in health status and function in patients with painful diabetic peripheral neuropathy</title>
		<author>
			<persName><surname>Dl Hoffman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">M</forename><surname>Sadosky</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Dukes</surname></persName>
		</author>
		<author>
			<persName><surname>Alvir</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Pain</title>
		<imprint>
			<biblScope unit="volume">149</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="194" to="201" />
			<date type="published" when="2010-05">May 2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">The computer expression recognition toolbox (cert)</title>
		<author>
			<persName><forename type="first">Gwen</forename><surname>Littlewort</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jacob</forename><surname>Whitehill</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Tingfan</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ian</forename><surname>Fasel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Mark</forename><surname>Frank</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Javier</forename><surname>Movellan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Marian</forename><surname>Bartlett</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Automatic Face &amp; Gesture Recognition and Workshops (FG 2011), 2011 IEEE International Conference on</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2011">2011</date>
			<biblScope unit="page" from="298" to="305" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">The consistency of facial expressions of pain: a comparison across modalities</title>
		<author>
			<persName><surname>Kenneth M Prkachin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Pain</title>
		<imprint>
			<biblScope unit="volume">51</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="297" to="306" />
			<date type="published" when="1992">1992</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Assessing pain by facial expression: facial expression as nexus</title>
		<author>
			<persName><surname>Kenneth M Prkachin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Pain Res Manag</title>
		<imprint>
			<biblScope unit="volume">14</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="53" to="58" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
