<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>” Journal on Multimodal User Interfaces</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Predicting Survey Responses of Social Constructs in a Dyad</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Bruno Abreu Calfa</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Peggy Wu</string-name>
          <email>Peggy.Wu@rtx.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mohammadamin Sanaei</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Stephen Gilbert</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andrew</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Radlbeck</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Brett Israelsen</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>SCOTTIE</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Systematic Communication</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Iowa State University</institution>
          ,
          <addr-line>527 Bissel Rd., Ames, Iowa</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Raytheon Technologies Research Center</institution>
          ,
          <addr-line>411 Silver Lane, East Hartford, Connecticut</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Social Computing</institution>
          ,
          <addr-line>Virtual Reality, Natural Language Analysis, Training</addr-line>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2022</year>
      </pub-date>
      <volume>4</volume>
      <issue>3</issue>
      <fpage>27</fpage>
      <lpage>29</lpage>
      <abstract>
        <p>Measuring social constructs such as engagement, rapport, and trust often rely heavily on surveys and behavioral observations. This paper describes a method to use features identified by psychology-based language analysis, combined with machine learning, to predict participant survey responses in a training context based on 120 dyad transcripts. The method analyzed data collected from subjects performing a circuit board training task within the project called Investigations and Evaluations. In this study, the collected data showed low utterance count and a lack of correlation between features and survey responses, suggesting that the context in which the interactions occurred may limit opportunities for interlocutors to manifest social behaviors verbally, which in turn affected the ability to use language analysis to predict subject perceptions of the interaction. However, the methodology appears sound.</p>
      </abstract>
      <kwd-group>
        <kwd>rapport</kwd>
        <kwd>perceived usability</kwd>
        <kwd>trust</kwd>
        <kwd>and mental</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        One modality in which humans exhibit social
behaviors is language. Linguistic categories [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]
have been used in diverse applications from
measuring emotional expression [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], to evaluating
team
dynamics
through
discourse
[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ],
and
identifying correlations between written student
self-introductions with course performance [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
This paper describes the use of Natural Language
Processing tools to examine transcripts between
trainer-student pairs in a project called Systematic
      </p>
      <sec id="sec-1-1">
        <title>Communication</title>
        <p>Objectives
and
Telecommunications Technology Investigations
and Evaluations (SCOTTIE). SCOTTIE’s goal is
to investigate the impact of the interaction media
on the effectiveness of achieving communication
objectives.</p>
        <p>
          The definition of communication
objectives is described in [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ].
        </p>
        <p>Briefly, the
communication objectives of interest include
copresence,
engagement,
virtual
embodiment,</p>
        <p>2020 Copyright for this paper by its authors. Use permitted under Creative
2.</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>Method</title>
      <p>The study protocol involved a scenario where
a trained confederate staff
member provided
scripted instructions on a circuit board repair task
to subjects. Subjects were assigned to one of three
conditions.</p>
      <p>Trainer-subject
interactions
were
either conducted through teleconference software
(i.e. Zoom), in a bespoke virtual reality based
environment (also called extended reality or XR),
or Face-to-Face (F2F) in-person visits with a
shared computer.</p>
      <p>
        All conditions used the same
circuit board simulator testbed, where the
trainersubject pair used screen share, controlled their
own avatars in the virtual environment, or shared
a physical screen, for the Zoom, XR and F2F
conditions respectively. The testbed and virtual
environment, called Circuit World, is software
created by the study staff as described in [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. At
the start of the trial, a research assistant explained
the purpose of the study and obtained informed
consent. The researcher then administered
pretrial surveys. Upon survey completion and other
introductory materials, the trainer entered the
session. The trainer provided subjects with
approximately 15 minutes of instruction on how
to repair a specific circuit board and invited
subjects to ask questions. The trainer then left the
session, and the researcher initiated the test
portion of the session, cuing the testbed for the
subject to repair a virtual circuit and complete a
multiple-choice quiz based on knowledge
conveyed during training. Subjects then
completed a post-session survey which contained
questions regarding their perception of the trainer
and the effectiveness of the communication
framed as the aforementioned communication
objectives. Only the transcript between the trainer
and the subject was used in the analysis. The
protocol was approved by Iowa State University’s
Internal Review Board (IRB). Participants were
recruited through Prolific, social media, and email
advertisements.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Feature Extraction from Transcript</title>
    </sec>
    <sec id="sec-4">
      <title>Data</title>
      <p>Each trainer-subject dyad transcript was
generated using Zoom’s auto transcription feature
and stored as VTT text files following the Web
Video Text Tracks format. Transcript files were
parsed to extract utterances by trainers and
subjects. In the F2F condition, some manual
transcript correction was needed due to the lack of
speaker diarization. The total number of
transcripts was 120, with 33, 50, and 37
transcripts from the F2F, Zoom, and XR
conditions respectively. Regarding the utterances
in each condition, as expected, the number of
utterances by trainers was reasonably consistent
across different conditions since confederate
trainers were following a script. There were
larger variations on the number of utterances by
subjects.</p>
      <p>The extraction of numerical features from
utterances in each transcript file was carried out
using the following natural language processing
(NLP) methodologies:</p>
      <p>
        Lexical and Semantic Similarity Analysis
using Word Embedding. The word embedding
approach is based on term-frequency times
inverse-document-frequency (tf-idf) [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], which is
a term weighting scheme calculated as
 - ( ,  ) =  ( ,  ) ⋅  ( ), where t is a term
and d is a document (i.e., utterance), tf(t,d) is a
matrix of counts of each uttered term in a
document, and  ( ) = log[(1 +  )/(1 +
 ( ))] + 1, where n is the number of documents
and df(t) is the number of documents in the
document set that contain term t. After embedding
the utterances by a trainer and a subject into their
respective tf-idf arrays, the cosine similarity score
[
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] is calculated as  ( ,  ) =   ⊤ , where x and
‖ ‖‖ ‖
y are tf-idf arrays and ‖ ‖ is the Euclidean norm.
      </p>
      <p>
        Linguistic Style Matching (LSM). LSM is a
technique in behavioral analytics to assess the
stylistic similarities in language use across groups
and individuals [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. The procedure measures the
degree of similarity between two individual’s
patterns of function word usage. Function words
consist of pronouns, articles, conjunctions,
prepositions, auxiliary verbs.
      </p>
      <p>
        Linguistic Inquiry and Word Count
(LIWC). LIWC is a word-counting software that
uses a dictionary containing words that belong to
over 80 linguistic, psychological, and topical
categories indicating various social, cognitive,
and affective processes [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. In this work, the
authors used the LIWC application programming
interface (API) offered by Receptiviti
(https://www.receptiviti.com/liwc).
      </p>
      <p>
        Valence-Aware Dictionary for Sentiment
Reasoning (VADER). VADER is a rule-based,
computational sentiment analysis method that
aims to measure the sentiments, evaluations,
attitudes, and emotions of a speaker/writer [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
The result of this analysis is a compound polarity
score for each utterance calculated as  /√ 2 +  ,
where x is the sum of sentiment scores of all the
words in the utterance and α is a normalization
parameter whose value is typically set to 15.
      </p>
      <p>The dataset of NLP extracted features
contained 197 columns, including LIWC features,
number of utterances by the subject, LSM
comparison between subject and trainer, VADER
compound polarity score for the subject, and
cosine similarity score between trainer and
subject.</p>
    </sec>
    <sec id="sec-5">
      <title>4. Building Predictive Models</title>
    </sec>
    <sec id="sec-6">
      <title>Communication Constructs for</title>
      <p>The overall data analytics workflow to
correlate communication constructs with trainer
and subject utterances is shown in Figure 1. It
comprised two pipelines: feature extraction (see
section above) and prediction. The prediction
pipeline used statistical and
Machine Learning
(ML) techniques for
data
preprocessing
and
regression analysis coupled with hyperparameter
tuning using Bayesian optimization.</p>
      <p>The target variables for the regression analysis
corresponded to survey responses related to each
communication construct. There were 9
highlevel constructs that were further divided into 12
targets based on the survey instruments used:
copresence,
engagement,
virtual
embodiment,
rapport, usability perceived ease, trust in the
trainer (general), trust (in
ability), trust (in
integrity),
workload
operating
trust
(in
benevolence),
mental
(general),
mental
workload
(in
the training
system), and
mental
workload (related to communication). Each target
variable had its own prediction pipeline. The
dataset of NLP extracted features was first merged
with the survey responses dataset (containing
target variables) on the subject identifier. Both
feature data and target values in the resulting
dataset were scaled between 0 and 1 (min-max
scaling), and then used in the prediction pipeline.</p>
      <p>
        The first step in the prediction pipeline was to
apply
a
variance-based
feature
reduction
procedure to the dataset, which removed all
lowvariance
features
according
to
a
variance
threshold value. The next step was to use a feature
selection approach to keep only the features that
had the n highest scores; the criterion was based
on
mutual information
between features and
target, using non-parametric estimation methods
based
on
entropy
estimation from
k-nearest
neighbors distances [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. The final step was to fit
a supervised learning regression model to predict
the target from the remaining features.
      </p>
      <sec id="sec-6-1">
        <title>Regression</title>
      </sec>
      <sec id="sec-6-2">
        <title>Algorithm</title>
        <p>Extreme</p>
        <p>
          Gradient
Boosting [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ]
        </p>
        <p>
          [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ]
        </p>
        <p>Gradient</p>
        <p>
          Boosting [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]
        </p>
        <p>The variance threshold, the value of n (number
of features with highest scores to be kept), and the
hyperparameters of a regression
model were
tuned simultaneously and systematically using a</p>
        <sec id="sec-6-2-1">
          <title>Bayesian optimization framework [13].</title>
          <p>The
optimization approach required the definition of a
search space describing the hyperparameters to be
tuned and their ranges, as well as an objective
function to guide the search. In this work, the
mean absolute error (MAE) is used as the
optimization criterion and calculated as 
1 ∑ |  −  ̂ |, where N was the number of records
=
or rows of the final dataset,   was the true target
valued of record  , and  ̂ was the predicted target
value of record  by the prediction pipeline.
search space of hyperparameters to be tuned by
the Bayesian Optimization Framework
Random Forest
n_estimators:{1000,1001,...,</p>
          <p>Hyperparameter Space
n_estimators:{1000,1001,...,
min_samples_split:[0.001,0.</p>
          <p>min_samples_leaf:</p>
          <p>[0.001,0.2]
n_estimators:{1000,1001,...,
optimize), more specifically, the cross-validated
search procedure over the hyperparameter search
space.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>5. Feature Importance with Shapley</title>
    </sec>
    <sec id="sec-8">
      <title>Values</title>
      <p>A prediction for a target variable can be
explained by assuming that each feature value of
the instance (i.e., survey record) is a "player" in a
game where the prediction is the payout. Shapley
values – a method from coalitional game theory –
tell us how to fairly distribute the "payout" among
the players [23]. Intuitively, Shapley values are
computed by carefully perturbing input features
and observing how changes to the input features
impact the final model prediction. The Shapley
value of a given feature is then calculated as the
average marginal contribution to the overall
model score.</p>
      <p>Mathematically, the Shapley value of player
(i.e., feature) i is calculated as
  ( ) =</p>
      <p>∑
 ⊆ \{ }
| |! ( − | | − 1)!
 !
( ( ∪ { }) −  ( ))
where  is a coalition/subset of players,  is the
number of players, and  (⋅) is a value function
that maps a subset of players to a real-valued
payout of the game. In other words, the Shapley
value was calculated by computing a weighted
average payout gain that player  provided when
included in all coalitions that excluded  .</p>
    </sec>
    <sec id="sec-9">
      <title>6. Result</title>
      <p>Figure 2 shows the mean absolute error (MAE)
of the best prediction pipeline on the test set for
each target. Note that the target values are scaled
between 0 and 1; therefore, on average, the
absolute error across all 12 targets varies between
11% and 22%, which does not exhibit a strong
correlation between NLP features extracted from
transcripts and overall communication construct
survey response scores. The results also show that
no single ML algorithm outperformed all others
for all targets.</p>
    </sec>
    <sec id="sec-10">
      <title>7. Discussion</title>
      <p>It is possible to interpret the relatively high
MAE as a lack of correlation between linguistic
features and survey responses. However, upon
further investigation of the data, the lack of
correlation may indeed be due to the nature of the
interaction within the designed scenario. A manual
examination of the interaction videos and
transcripts revealed that trainer-subject pairs
typically greeted each other with one sentence and
proceeded to the task of training without further
socializing. While the trainer spoke from the
script, and was therefore extremely consistent
across conditions, most utterances from the subject
were one- or two-word sentences such as “yes”,
“no”, “uh-huh”, or “I understand.” In examining
the video recordings, throughout the
approximately 15 minutes of training, subjects
appeared to be attentive and potentially
cognitively loaded with the task of listening and
absorbing the instruction. Questions from the
subject were often clarifying questions or asking
the trainer to repeat. After the training session, the
trainer and subject did not have the opportunity for
any unplanned incidental conversation. It appears
that the subject simply did not have the
opportunity or cognitive resources to exhibit
verbal behaviors hinting at their level of
engagement, rapport, trust, or sense of co-presence
with the trainer, or perceived workload.
Interactions that are lower in cognitive workload
or are richer in social exchange may provide more
opportunities for linguistic markers to manifest. In
addition to the lack linguistic manifestations, the
authors previously reported significant differences
in objective task performance but no significant
differences in social constructs such as those
reported above [24] between conditions. One
possible interpretation is that social interactions
occur during “off-duty” time gaps between
sessions, whether in-person, over video, or in
virtual environments. When training sessions are
highly controlled and time constrained such as our
design, participants are “on-duty” and not
exhibiting social behaviors. If social behaviors are
desired, such as in newly formed teams, one
recommendation may be to build in time gaps to
afford such incidental or informal interactions to
occur regardless of the communication medium.</p>
    </sec>
    <sec id="sec-11">
      <title>8. Acknowledgements</title>
      <p>The information, data, or work presented
herein was funded in part by the Advanced
Research Projects Agency-Energy (ARPA-E),
U.S. Department of Energy, under Award
Number DE-AR0001097. The views and opinions
of authors expressed herein do not necessarily
state or reflect those of the United States
Government or any agency thereof. The authors
would also like to thank Jack Geddes and Jennifer
Glista from Receptiviti.com for their guidance
and support.
9. References</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Y. R.</given-names>
            <surname>Tausczik</surname>
          </string-name>
          and
          <string-name>
            <given-names>J. W.</given-names>
            <surname>Pennebaker</surname>
          </string-name>
          , “
          <article-title>The psychological meaning of words: LIWC and computerized text analysis methods</article-title>
          ,
          <source>” Journal of language and social psychology</source>
          , vol.
          <volume>29</volume>
          , no.
          <issue>1</issue>
          , pp.
          <fpage>24</fpage>
          -
          <lpage>54</lpage>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J. H.</given-names>
            <surname>Kahn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. M.</given-names>
            <surname>Tobin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. E.</given-names>
            <surname>Massey</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J. A.</given-names>
            <surname>Anderson</surname>
          </string-name>
          , “
          <article-title>Measuring emotional expression with the Linguistic Inquiry</article-title>
          and Word Count,”
          <article-title>The American journal of psychology</article-title>
          , vol.
          <volume>120</volume>
          , no.
          <issue>2</issue>
          , pp.
          <fpage>263</fpage>
          -
          <lpage>286</lpage>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>C.</given-names>
            <surname>Miller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Rye</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Wu</surname>
          </string-name>
          , S. SchmerGalunder, and T. Ott, “
          <article-title>Team psychosocial assessment via discourse analysis: Power and comfort</article-title>
          /routine,” in International Conference on Social Computing,
          <string-name>
            <surname>Behavioral-Cultural Modeling</surname>
          </string-name>
          , and
          <string-name>
            <surname>Prediction</surname>
          </string-name>
          ,
          <year>2014</year>
          , pp.
          <fpage>309</fpage>
          -
          <lpage>316</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>R. L.</given-names>
            <surname>Robinson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Navea</surname>
          </string-name>
          , and W. Ickes, “
          <article-title>Predicting final course performance from students' written self-introductions: A LIWC analysis</article-title>
          ,
          <source>” Journal of Language and Social Psychology</source>
          , vol.
          <volume>32</volume>
          , no.
          <issue>4</issue>
          , pp.
          <fpage>469</fpage>
          -
          <lpage>479</lpage>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>R. E.</given-names>
            <surname>Dianiska</surname>
          </string-name>
          et al.,
          <article-title>“Do You Need to Travel? Mapping Face-to-Face Communication Objectives to Technology Affordances,”</article-title>
          <source>Proceedings of the Human Factors and Ergonomics Society Annual Meeting</source>
          , vol.
          <volume>64</volume>
          , no.
          <issue>1</issue>
          , pp.
          <fpage>1069</fpage>
          -
          <lpage>1073</lpage>
          , Dec.
          <year>2020</year>
          , doi: 10.1177/1071181320641256.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>J.</given-names>
            <surname>Rozell</surname>
          </string-name>
          et al., “
          <article-title>Circuit world: a multiplayer VE for researching engineering learning</article-title>
          ,” in
          <source>2021 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>773</fpage>
          -
          <lpage>773</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>R.</given-names>
            <surname>Baeza-Yates</surname>
          </string-name>
          and
          <string-name>
            <given-names>B.</given-names>
            <surname>Ribeiro-Neto</surname>
          </string-name>
          ,
          <article-title>Modern Information Retrieval</article-title>
          .
          <source>Addison Wesley</source>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>C. D.</given-names>
            <surname>Manning</surname>
          </string-name>
          , Introduction to information retrieval.
          <source>Syngress Publishing</source>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>M. E.</given-names>
            <surname>Ireland</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. B.</given-names>
            <surname>Slatcher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. W.</given-names>
            <surname>Eastwick</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. E.</given-names>
            <surname>Scissors</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. J.</given-names>
            <surname>Finkel</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J. W.</given-names>
            <surname>Pennebaker</surname>
          </string-name>
          , “
          <article-title>Language style matching predicts relationship initiation and stability,” Psychological science</article-title>
          , vol.
          <volume>22</volume>
          , no.
          <issue>1</issue>
          , pp.
          <fpage>39</fpage>
          -
          <lpage>44</lpage>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>J.</given-names>
            <surname>Pennebaker</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Francis</surname>
          </string-name>
          , Linguistic Inquiry and
          <string-name>
            <given-names>Word</given-names>
            <surname>Count</surname>
          </string-name>
          . Lawrence Erlbaum Associates, Inc.,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>C.</given-names>
            <surname>Hutto</surname>
          </string-name>
          and E. Gilbert, “
          <article-title>Vader: A parsimonious rule-based model for sentiment analysis of social media text</article-title>
          ,”
          <source>in Proceedings of the international AAAI conference on web and social media</source>
          ,
          <year>2014</year>
          , vol.
          <volume>8</volume>
          , no.
          <issue>1</issue>
          , pp.
          <fpage>216</fpage>
          -
          <lpage>225</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>A.</given-names>
            <surname>Kraskov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Stögbauer</surname>
          </string-name>
          , and
          <string-name>
            <given-names>P.</given-names>
            <surname>Grassberger</surname>
          </string-name>
          , “Estimating mutual information,” Physical review
          <string-name>
            <surname>E</surname>
          </string-name>
          , vol.
          <volume>69</volume>
          , no.
          <issue>6</issue>
          , p.
          <fpage>066138</fpage>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>E.</given-names>
            <surname>Brochu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. M.</given-names>
            <surname>Cora</surname>
          </string-name>
          , and N. De Freitas, “
          <article-title>A tutorial on Bayesian optimization of expensive cost functions, with application to active user modeling and hierarchical reinforcement learning</article-title>
          ,
          <source>” arXiv preprint arXiv:1012.2599</source>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>T.</given-names>
            <surname>Chen</surname>
          </string-name>
          and
          <string-name>
            <given-names>C.</given-names>
            <surname>Guestrin</surname>
          </string-name>
          , “
          <article-title>Xgboost: A scalable tree boosting system,” in Proceedings of the 22nd acm sigkdd international conference on knowledge discovery and data mining</article-title>
          ,
          <year>2016</year>
          , pp.
          <fpage>785</fpage>
          -
          <lpage>794</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>L.</given-names>
            <surname>Breiman</surname>
          </string-name>
          , “Random forests,”
          <article-title>Machine learning</article-title>
          , vol.
          <volume>45</volume>
          , no.
          <issue>1</issue>
          , pp.
          <fpage>5</fpage>
          -
          <lpage>32</lpage>
          ,
          <year>2001</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>J. H.</given-names>
            <surname>Friedman</surname>
          </string-name>
          , “
          <article-title>Greedy function approximation: a gradient boosting machine,”</article-title>
          <source>Annals of statistics</source>
          , pp.
          <fpage>1189</fpage>
          -
          <lpage>1232</lpage>
          ,
          <year>2001</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>P.</given-names>
            <surname>Geurts</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ernst</surname>
          </string-name>
          , and L. Wehenkel, “
          <article-title>Extremely randomized trees</article-title>
          ,
          <source>” Machine learning</source>
          , vol.
          <volume>63</volume>
          , no.
          <issue>1</issue>
          , pp.
          <fpage>3</fpage>
          -
          <lpage>42</lpage>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>G. E.</given-names>
            <surname>Hinton</surname>
          </string-name>
          , “
          <article-title>Connectionist learning procedures,” in Machine learning</article-title>
          ,
          <source>Elsevier</source>
          ,
          <year>1990</year>
          , pp.
          <fpage>555</fpage>
          -
          <lpage>610</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>C.-C. Chang</surname>
          </string-name>
          , “
          <article-title>' LIBSVM: a library for support vector machines</article-title>
          ,
          <source>' ACM Transactions on Intelligent Systems and Technology</source>
          ,
          <volume>2</volume>
          :
          <issue>27</issue>
          :
          <fpage>1</fpage>
          -
          <lpage>27</lpage>
          :
          <fpage>27</fpage>
          ,
          <year>2011</year>
          ,” http://www. csie. ntu. edu. tw/∼ cjlin/libsvm, vol.
          <volume>2</volume>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>J.</given-names>
            <surname>Goldberger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. E.</given-names>
            <surname>Hinton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Roweis</surname>
          </string-name>
          , and
          <string-name>
            <given-names>R. R.</given-names>
            <surname>Salakhutdinov</surname>
          </string-name>
          , “
          <source>Neighbourhood Components Analysis,” in Advances in Neural Information Processing Systems</source>
          ,
          <year>2004</year>
          , vol.
          <volume>17</volume>
          . Accessed: Aug.
          <volume>30</volume>
          ,
          <year>2022</year>
          . [Online]. Available: https://papers.nips.cc/paper/2004/hash/42fe8808 12925e520249e808937738d2-
          <fpage>Abstract</fpage>
          .html
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>L.</given-names>
            <surname>Bottou</surname>
          </string-name>
          and
          <string-name>
            <given-names>O.</given-names>
            <surname>Bousquet</surname>
          </string-name>
          , “
          <article-title>The tradeoffs of large scale learning</article-title>
          ,
          <source>” Advances in neural information processing systems</source>
          , vol.
          <volume>20</volume>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>