<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Simulating Human Detection
of Phishing Websites: An Investigation into the Applicabil-
ity of the ACT-R Cognitive Behaviour Architecture Model.
In</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>On Interactive Machine Learning and the Potential of Cognitive Feedback</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>Chris J. Michael, Dina Acklin, and Jaelle Scheuerman U.S. Naval Research Laboratory 1005 Balch Blvd, Code 7343 Stennis Space Center</institution>
          ,
          <addr-line>Mississippi 39529.</addr-line>
          <country country="US">U.S.A</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>The Emergence of Interactive Machine Learning</institution>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2017</year>
      </pub-date>
      <volume>3</volume>
      <fpage>1</fpage>
      <lpage>8</lpage>
      <abstract>
        <p>In order to increase productivity, capability, and data exploitation, numerous defense applications are experiencing an integration of state-of-the-art machine learning and AI into their architectures. Particular to defense applications, having a human analyst in the loop is of high interest due to quality control, accountability, and complex subject matter expertise not readily automated or replicated by AI. However, many applications are suffering from a very slow transition. This may be in large part due to lack of trust, usability, and productivity, especially when adapting to unforeseen classes and changes in mission context. Interactive machine learning is a newly emerging field in which machine learning implementations are trained, optimized, evaluated, and exploited through an intuitive human-computer interface. In this paper, we introduce interactive machine learning and explain its advantages and limitations within the context of defense applications. Furthermore, we address several of the shortcomings of interactive machine learning by discussing how cognitive feedback may inform features, data, and results in the state of the art. We define the three techniques by which cognitive feedback may be employed: self reporting, implicit cognitive feedback, and modeled cognitive feedback. The advantages and disadvantages of each technique are discussed.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        This will certify that all author(s) of the above article/paper are
employees of the U.S. Government and performed this work as part of
their employment, and that the article/paper is therefore not subject
to U.S. copyright protection. No copyright. Use permitted under
Creative Commons License Attribution 4.0 International (CC BY
4.0). In: Proceedings of AAAI Symposium on the 2nd Workshop
on Deep Models and Artificial Intelligence for Defense
Applications: Potentials, Theories, Practices, Tools, and Risks, November
11-12, 2020, Virtual, published at http://ceur-ws.org
the system. Though undoubtedly useful for commercial
bigdata problems, there are many scenarios – especially in
defense – where applying AML falls short in practice. For
instance, applications at the tactical edge may suffer from
smaller quantities of labeled examples for training.
Moreover, classifiers may struggle to adapt to changes in data
context quickly enough to be considered viable by an analyst,
particularly in scenarios where the mission demands quick
turn-around time. Many of these issues may be mitigated by
emerging implementations of the interactive machine
learning (IML) paradigm, which capitalizes on human input in
order to improve machine learning implementations
        <xref ref-type="bibr" rid="ref11">(Fails
and Olsen Jr 2003)</xref>
        . Unlike approaches that leverage AML,
IML implementations allow for classifiers to very quickly
train and apply newly discovered information with the help
of a human subject-matter expert, which we refer to in this
article as the analyst.
      </p>
      <p>
        In general, IML may be described as a machine learning
implementation where one or more analysts iteratively
improve a model for automation by manipulating an interface
that is tightly coupled to the desired task at hand. There are
four main components to any IML implementation. The first
component is the data associated with the task. Examples
of such data include remotely sensed imagery, textual
information such as reports, and spatiotemporal tracks of moving
objects. The second component, referred to in this study as
the machine, is the mathematical model that tries to estimate
or automate the desired task. Ostensibly, this can be seen as
a black-box, but we will discuss the properties of a
successful IML classifier later in the article. The third component of
IML is the Human-Computer Interface (HCI). The HCI may
be as conventional as software receiving input through a
keyboard and mouse, which is what we assume in this article, or
as specialized as vehicle controls, immersive environments,
and brain interfaces. The application is designed to allow
immediate and intuitive presentation of the machine’s
classification on a manageable set of data. This data is then either
confirmed or manipulated to be correct by the analyst, who
is the last but most important component of an IML system.
In this article, we discuss IML within the context of
improving productivity and decision making for an analyst with
a very specific task that requires subject-matter expertise.
Though, as exemplified above, IML may be deployed in a
wide variety of ways, we feel that deployment in this context
has the greatest potential for impact in defense applications.
There are several studies that provide excellent perspectives
of the current state of the art in IML outside of this scope
        <xref ref-type="bibr" rid="ref10 ref16 ref20 ref37 ref43 ref47">(Dudley and Kristensson 2018; Wu, Weld, and Heer 2019;
Robert et al. 2016)</xref>
        .
      </p>
      <p>A common architecture for IML implementations is
shown in Figure 1. The data on which the analyst must
perform a task may either be completely available in a database
or sequentially available as a stream. Active learning may be
used to pull the most effective data points from this database
for labeling, as will be discussed in the next section. A
machine for predicting the data is then used to present guesses
for the task at hand to the analyst. The analyst must verify
each of these guesses and correct any mistakes via the HCI.
Once the verification step is completed for the current
iteration, the machine will immediately learn from the
corrections and/or confirmations. The process will then repeat by
the machine gathering data examples and presenting guesses
to the user once again. When the time comes for the analyst
to leave their duty station, the machine model may optimize
on the data that has been labeled in order to maximize its
accuracy. This way, the most effective machine will be
available once the analyst returns to duty. It is important to note
that the machine may be deployed as a centralized
generalpurpose classifier that combines the work done by multiple
analysts, or it may be deployed locally to be custimized
towards the individual analyst.</p>
      <p>
        The focus of this article is to introduce IML within
the context of analyst-driven applications relevant to
defense while highlighting research gaps, the most important
of which involves incorporation of cognitive feedback. We
choose not to discuss manual model interactions such as
feature selection
        <xref ref-type="bibr" rid="ref36">(Raghavan, Madani, and Jones 2006)</xref>
        or
model selection
        <xref ref-type="bibr" rid="ref40">(Talbot et al. 2009)</xref>
        , which are processes
whereby analysts directly optimize machine models. Rather,
we choose to present implementations that can be used
effectively by an analyst who is a subject matter expert for the
task at hand and not knowledgeable in machine learning or
statistical theory. Defense analysts hold invaluable
subjectmatter expertise for the mission, and it is unreasonable to
assume that they must learn or worry about data-scientific
concepts. Because intuitive HCIs may be designed to be
congruent to their task, IML has great potential to leverage the
power of modern-day ML while not burdening the analyst
with parameter tuning, data curation, or any of the other
burdens implicit to AML.
      </p>
      <p>The next section will describe three examples of IML
implementations that highlight the current state of the art. The
section that follows will iterate through several advantages,
shortcomings, and gaps in the state of the art. In the
penultimate section, we specify the ways in which cognitive
feedback may be used to address the shortcomings and gaps of
IML with respect to defense applications. Finally, we
conclude with commentary on prospects for future research.</p>
    </sec>
    <sec id="sec-2">
      <title>Interactive Machine Learning in Action</title>
      <p>In order to frame a more detailed discussion of IML, we
now describe several IML implementations that have been
presented in peer-reviewed literature. We specifically choose
three application areas that are analyst-driven: region
digitization, textual translation, and video annotation. These
examples demonstrate the potential for IML to improve both
the machine performance and the user experience with
autonomy.</p>
      <p>
        Geographic region digitization is a highly demanded yet
arduous task whereby regions such as bodies of water and
other land cover are digitized from remotely sensed images,
usually within a Geographic Information System
        <xref ref-type="bibr" rid="ref16 ref20 ref43">(Hossain
and Chen 2019)</xref>
        . Once digitized, regions may be represented
in mapping products for geospatial situational awareness,
climate-level studies, land surveys, and many other
applications. Although numerous AML approaches to region
digitization have been presented in the literature, they are not
widely adopted in practice. This is most likely due to the
all-or-nothing yield of AML approaches: If the machine
incorrectly digitizes a region, it may be more burdensome for
an analyst to correct than to start from scratch. Therefore, an
analyst may prefer to digitize manually to circumvent
frustration and presumably lower their workload. In order to
address these shortcomings, an IML implementation for region
digitization, named the Geospatial Region Application
Interface Toolkit (GRAIT), is presented as a human-machine
team application
        <xref ref-type="bibr" rid="ref33">(Michael et al. 2019)</xref>
        . The authors address
the all-or-nothing approach to region digitization with an
IML implementation where a region is digitized iteratively.
In each iteration, the machine guesses the placement of a
certain number of vertices of the contour and presents them
to the analyst for verification. For each vertex presented,
the analyst may either correct its placement by clicking and
dragging it to an appropriate location or simply confirm its
correct placement by not interacting with it. The analyst
indicates via button press when all the vertices of the current
iteration are corrected or confirmed. The machine will then
train on the finalized vertex locations, and the process will
continue until the region is completely digitized. In order to
prevent inducing too high of a cognitive load on the
analyst, an uncertainty model is used to estimate the
probability of incorrect vertex placement and limit each iteration to
around 2 incorrectly placed vertices. Results show that with
no prior training data, the IML implementation accurately
places 84% of vertices correctly in 4 separate image sets of
4 images each.
      </p>
      <p>
        Another area where IML approaches show promise is that
of textual language translation, commonly referred to as
machine translation. While bodies of work in this field attempt
to replace human translators with machine models, many of
which are AML implementations
        <xref ref-type="bibr" rid="ref23">(Koehn 2009)</xref>
        , the current
state of the art is far from perfect. As with region
digitization, fully-automatic approaches may hinder rather than
help the performance of a translator at times when too many
mistranslated words may induce excessive cognitive load.
Because of this, many approaches to machine translation
are realized through a human-machine team. An IML
approach to machine translation aims to remedy these issues by
implementing iterative learning and modeling the
informativeness of each machine translation at a fine-grained level
        <xref ref-type="bibr" rid="ref13">(Gonza´lez-Rubio and Casacuberta 2014)</xref>
        . In this approach,
an initial guess of a sentence translation is given to the user
Active
Learning
      </p>
      <p>Guess
Online
Learning
Refine</p>
      <p>Interface
(HCI)</p>
      <p>Analyst
Verify
Correct</p>
      <p>Optimization
based on a metric of informativeness. The user will then
make corrections to the guess by changing the first
incorrect letter of the translation. The machine in turn suggests
a new translation under this assumption. This process
continues, with the machine immediately training on corrected
data for future translations. Results show that employing this
IML-based method produces twice the translation quality, a
metric specific to machine translation, per user interaction
over AML approaches.</p>
      <p>
        Lastly, IML implementations have emerged for the
difficult task of video annotations, where the amount of data
generated per day has far surpassed the ability of analysts to
inspect. When successful, annotated video allows for critical
advantages such as the ability to search for events, quantify
behavioral analytics, and study natural phenomena. Though
many AML approaches to video analytics exist, they are
typically tied to certain features of interest within some
constrained context
        <xref ref-type="bibr" rid="ref2">(Ananthanarayanan et al. 2017)</xref>
        . In cases
where context may change and the features of interest are
unknown, AML implementations for automatic video
annotation may be rendered incorrect or infeasible. This is
especially true in cases where context has changed or features of
interest are unknown beforehand. An IML implementation
of video annotation named Janelia Automatic Animal
Behavior Annotator (JAABA) demonstrates a semi-automatic
approach to assess animal behavior
        <xref ref-type="bibr" rid="ref21">(Kabra et al. 2013)</xref>
        .
JAABA allows for a user to annotate a video frame with an
arbitrary label, for instance jump. Then, using trajectory
information extracted from the video, the machine trains on
the given label and presents classification results both at the
level of the current video and a database of numerous
animal videos. The machine also provides confidence levels
for each classification to guide further labeling by the user.
This process is repeated iteratively until an ideal classifier
is attained. JAABA was used to create the first ML-driven
behavior classifier over a diverse set of animals.
      </p>
      <p>With these three examples in mind, a more detailed
explanation of the advantages, limitations, and gaps of IML will
follow.</p>
    </sec>
    <sec id="sec-3">
      <title>Advantages, Shortcomings, and Gaps</title>
      <sec id="sec-3-1">
        <title>Advantages</title>
        <p>
          The advantages of IML approaches directly address many
of the shortcomings that defense applications exhibit when
utilizing ML. Numerous defense applications suffer from
a shortage of labeled training examples due to a lack of
crowd sourcing and the ever changing state of platform
technologies among other reasons. As such, deep models
relying on large amounts of labeled examples cannot be
adequately trained. IML addresses the shortage of training data
by providing an interface that allows for incorrect
classifications to be immediately corrected and integrated into the
machine model. In fact, several IML implementations may
work well with no prior labeled data, which is usually
referred to as the cold start problem
          <xref ref-type="bibr" rid="ref13 ref27">(Lika, Kolomvatsos, and
Hadjiefthymiades 2014)</xref>
          . Additionally, the HCI allows for
correction through an intuitive interface that potentially
reduces the burden of data labeling. This allows an analyst to
leverage their current subject-matter expertise – that of the
application and data context – and circumvents the need to
play the role of data scientist.
        </p>
        <p>
          Defense problems must be very adaptable to context
changes from one region of interest to the next. In order to
accommodate this, any autonomy must immediately adapt
to such changes at the pace of the analyst. Therefore, IML
implementations typically apply active learning and online
learning techniques in order to improve effectiveness. Active
learning research entails the study of uncertainty or
similarity metrics in order to develop a mathematical
understanding of the likelihood that a machine will classify future data
points correctly
          <xref ref-type="bibr" rid="ref35">(Quionero-Candela et al. 2009)</xref>
          . The field
of online machine learning involves models that may train
in stride to adapt to new situations quickly while optimizing
exploration vs. exploitation
          <xref ref-type="bibr" rid="ref8">(Bottou 1998)</xref>
          .
        </p>
        <p>Problems related to defense must sometimes be deployed
at the tactical edge. In such situations, computational
resources and downtime may be scarce. IML directly
addresses this problem, since most IML implementations are
meant to be deployed on desktop computers. In all three
examples of IML presented in the previous section, online and
active learning strategies are employed to iteratively build
high-performance classifiers. Active learning is also used to
gage the load of examples presented to the user, both by
correlating uncertainty to the probability of an incorrect
classification and by providing a priority for the analyst to
manage their own work flow. Both GRAIT and JAABA support
cold-start cases.</p>
      </sec>
      <sec id="sec-3-2">
        <title>Shortcomings</title>
        <p>
          Perhaps the most obvious shortcoming if IML is that the
HCI and machine implementation must be tightly coupled
to a specific application. This entails much more effort in
the development of applications, since they must be built
and studied uniquely towards an explicit work flow. This
differs greatly from AML approaches, where for the most
part implementations are general-purpose and specificity is
implied through parameterization and classes for labeling.
Studies define a general-purpose methodology for HCI, but
this research is young and remains mostly theoretical in
nature
          <xref ref-type="bibr" rid="ref16 ref20 ref43">(Meza Mart´ınez and Maedche 2019)</xref>
          .
        </p>
        <p>
          Deep models of machine learning exhibit very
impressive results relating to throughput of data and classification
times. IML implementations currently lag behind in these
results. This is in part due to the nature of online machine
learning; namely, the need to have tight classification and
training cycles. However, research is trending more towards
online and active learning problems, and IML-inspired
classifiers with competent performance are emerging
          <xref ref-type="bibr" rid="ref26 ref31 ref41 ref45">(Langford,
Li, and Strehl 2007; Lu, Shi, and Jia 2013)</xref>
          .
        </p>
        <p>
          A further issue with IML is that overfitting may occur
more frequently since data is generally labeled iteratively.
Overfitting occurs when prior training data causes the model
to correlate too tightly to features that do not justify the
desired outcome. For example, one of the geographic sites
in the GRAIT study is Johnson Lake, WA. The first three
images show the shoreline in roughly the same location.
The fourth image shows the lake with a receded shoreline.
Though the shoreline may be spotted by an analyst clearly in
the fourth image, the classifier overfit to spaital features and
thus incorrectly identified the shoreline. This also caused
the uncertainty calculations for the image to be undershot.
AML approaches to overfitting typically require
optimizing machine parameters or adding diversity to datasets, both
of which typically require large amounts of computation
and thus long turnaround times not conducive to
successful IML implementations. Therefore, reinforcement
metalearning, whereby active learning implementations are
informed by corrections via specialized ML implementations,
may be employed to adapt quickly to situations where
overfitting is inevitable
          <xref ref-type="bibr" rid="ref5">(Bachman, Sordoni, and Trischler 2017)</xref>
          .
        </p>
      </sec>
      <sec id="sec-3-3">
        <title>The Cognitive Gap</title>
        <p>
          Although frequently mentioned as a future direction of
study, perhaps the largest identified gap in IML research is
the lack of formalization and quantification of cognitive
implications from the analyst. For instance, the IML machine
translation study
          <xref ref-type="bibr" rid="ref13">(Gonza´lez-Rubio and Casacuberta 2014)</xref>
          mentions specifically that the applied technique lessens the
cognitive load of the translator by utilizing cost-sensitive
metrics such as informativeness. However, the study does
not perform any human-factors research to back support this
claim, though it is mentioned as future work. As another
example, the study presenting GRAIT uses mathematically
modeled uncertainty calculations to meter the workload at
each iteration. Though it is shown statistically that these
uncertainty calculations correlate to the probability a vertex
is placed correctly, results focus more on vertex placement
accuracy and do not consider multiple load levels (e.g. the
number of expected incorrect vertices is set to two for the
entire study). Human factors research is also slated as future
work. Both of these studies appreciate that there must be
thresholds of cognitive load taken into account by the IML
system for a successful implementation, but it is apparent
that human-factors research is inevitable.
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>The Implications of Cognitive Feedback</title>
      <p>
        Due to its interactive nature, IML most certainly is a
humanin-the-loop endeavor. Several studies have highlighted
difficulties that may arise from trust, safety, and quality
        <xref ref-type="bibr" rid="ref10 ref12 ref15 ref16 ref20 ref43 ref44">(Dudley
and Kristensson 2018; Groce et al. 2013; Gillies et al. 2016;
Turchetta, Berkenkamp, and Krause 2019)</xref>
        . This section is
devoted to discussing the potential of researching and
integrating models of human cognition as feedback for IML,
which is not often mentioned in the state of the art. We
also make the argument that cognitive feedback directly
addresses the shortcomings of IML. The topic of cognitive
feedback is especially useful for defense-related problems,
where trust, safety, and quality of ML implementations is a
prerequisite for adoption. Without analyst-driven cognitive
feedback, an IML system can very quickly fall flat, which is
illustrated in the following region digitization example.
      </p>
      <p>Consider the analyst using GRAIT to digitize the fourth
image of Johnson Lake as explained in the previous
section. Recall that the machine is overfit, and thus its model
for uncertainty is undershot. Because of this, the machine
places 10 vertices, 8 of which are incorrectly placed. If the
analyst continues, they will spend more time correcting the
misplaced vertices than manually digitizing the lake without
the help of the machine.</p>
      <p>This example is simple, but it highlights one of the
detrimental problems of IML implementations: Overfitting is
inevitable, and it can induce, rather than relieve, cognitive
load. As mentioned previously, reinforcement learning may
be used to augment the uncertainty or similarity model based
on the number of corrections the user has to make in any
iteration. However, convergence of such a technique would
involve the user making excessive corrections in order to
inform the model in this example. Unlike AML, the
uncertainty and workload involved with IML data must be
somehow informed by the analyst.</p>
      <p>Figure 2 shows several situations exemplifying various
levels of cognitive load when an analyst uses GRAIT to
annotate some region of interest. In the first example, the
machine is very accurate but offers too few vertices for the
analyst to verify. In this situation, the analyst is impeded by
an overshot cognitive load. The analyst must work at the
slow pace of the IML implementation, which not only
reduces their productivity but may also reduce their attention
and engagement. The second example shows the ideal
situation where GRAIT correctly manages the cognitive load of
the analyst. The analyst is expected to be engaged and
productive. The last situation shows an example of the IML
implementation undershooting the cognitive load. This causes
the analyst to become overwhelmed and possibly confused,
slowing their productivity and causing frustration.
?!?</p>
      <p>Incorporation of cognitive load is necessary to avoid the
pitfall of bad cognitive load estimation based on analysis of
data alone. For instance, consider an augmentation to the
third GRAIT example in the figure by providing the user
with a survey at each iteration. The survey will occur before
correction and simply ask, “Is this workload too little, too
much, or fine?” In this particular situation, the analyst will
inform the machine that the workload is too much to handle,
and the machine may modify its uncertainty model
accordingly (e.g. by adjusting weighting or performing best-fit
optimization to prior iterations). This very simple solution
illustrates how cognitive feedback may enable better IML for
many applications, but this concept may be taken further. In
order to promote discussion and research of the possibilities
and implications of this concept, we now present a
taxonomy for cognitive feedback to inform IML.</p>
      <p>
        Self-reported cognitive feedback is gathered by surveys
eliciting cognitive feedback from the user. An example of
such a survey is the standard NASA-TLX, which allows
a user to report on the general experienced workload of a
particular task
        <xref ref-type="bibr" rid="ref18">(Hart and Staveland 1988)</xref>
        . This could be
gathered offline during human factors evaluation or online
through an interface for self reporting within the HCI. The
main advantage of online self reporting cognitive load is the
simplicity to collect feedback within the HCI.
Implementation of simple interventions, such as providing buttons for
when a workload is too heavy or too light, are trivial.
However, this approach may be imprecise in complex user
environments because sub-components of a task may
differentially contribute to workload. In these situations,
interventions may be too simplistic or induce load on an analyst.
      </p>
      <p>
        Until now, we’ve discussed the implications of self
reporting on cognitive load, but this technique may provide insight
into more than just the analyst’s ideal workload. The field
of explainable artificial intelligence involves expressing the
machine’s decision making to a human user
        <xref ref-type="bibr" rid="ref16 ref20 ref43">(Gunning and
Aha 2019)</xref>
        . If a model for explainability is feasible, then the
user may communicate cognitive information relating to
features as feedback to the model
        <xref ref-type="bibr" rid="ref16 ref20 ref43">(Teso and Kersting 2019)</xref>
        .
Relating back to the example above, the machine may explain
its decisions by stating “I believe that historic position of the
shoreline is very important.” The user may then augment the
belief by stating “The historic position is not as important as
color,” and the machine may then optimize its classifier and
uncertainty calculation based on this statement.
      </p>
      <p>
        As opposed to surveying a user, implicit cognitive
feedback may be collected in real time while analysts interact
with the HCI during closed experimentation. Implicit
cognitive feedback involves collecting physiological data in order
to infer cognitive states in a manner that is continuous,
objective, and occurs in real time. For example, because
pupillary responses are reflective of nervous activity, pupil
dilation may act as a proxy for measuring task-induced
cognitive processes. As such, increases in pupil diameter may
be indicative of high cognitive load, attentional processing,
and decision making
        <xref ref-type="bibr" rid="ref17 ref19 ref22">(Hess and Polt 1964; Kahneman 1973;
Hahnemann and Beatty 1967)</xref>
        whereas decreases may reflect
fatigue
        <xref ref-type="bibr" rid="ref29">(Lowenstein, Feinberg, and Loewenfeld 1963)</xref>
        . This
data may then be correlated with self-reporting to define
various states of cognitive load. Examples of such
biofeedback include readings of skin conductance, heart rate,
pupilometry, and electroencephalogram (EEG). Often, multiple
physiological measures will be assessed to determine
workload and inform adaptive algorithms, in essence creating
user models that dynamically adjust to support user needs.
For example, such physiological elements were examined to
monitor the workload of operators while performing UAV
piloting tasks of different levels
        <xref ref-type="bibr" rid="ref41 ref45">(Wilson and Russell 2007)</xref>
        .
The physiological signals were used as features to train
a neural network to classify workload. Another approach
of implicit cognitive feedback is to incorporate cognitive
cues as features in the machine learning algorithm
(Rosenfeld et al. 2012). For example, in a recent choice
competition, researchers incorporated cognitive features derived
from behavior into a random forest algorithm. They found
that this approach significantly outperformed other ML
approaches that did not incorporate cognitive features
        <xref ref-type="bibr" rid="ref34">(Plonsky et al. 2017)</xref>
        . A recent study has explored how collecting
and applying cognitive cues as features improves
reinforcement learning algorithms for playing video games
        <xref ref-type="bibr" rid="ref48">(Zhang
et al. 2019)</xref>
        . In summary, implicit cognitive feedback has
the potential to improve IML implementations by gathering
data in closed experimentation to inform cognitive load,
uncertainty/similarity measurements, and inform the machine
with features of interest related to a specific task.
      </p>
      <p>Implicit cognitive feedback may provide invaluable
in</p>
      <sec id="sec-4-1">
        <title>Implicit</title>
      </sec>
      <sec id="sec-4-2">
        <title>Modeled Utilization of a cognitive model in the loop. Table 1: Taxonomy of Cognitive Feedback for Interactive Machine Learning</title>
        <p>
          sight to IML implementations, but the disadvantage lies in
the fact that closed experimentation is often necessary to
collect biofeedback, control levels of tasking, and survey users
of the HCI with respect to a particular application.
Additionally, the cognitive state of the user may be more dynamic for
some applications than others. In these situations, modeled
cognitive feedback may provide cognitive feedback based
on models of user interaction with the HCI. For example,
simulating human behavior using a computational cognitive
model is another potential method to provide feedback to an
IML system. Models of cognition and decision making have
been used to simulate human interactions with interfaces in
military contexts
          <xref ref-type="bibr" rid="ref6">(Blasch et al. 2011)</xref>
          . Cognitive
architectures represent a modeling paradigm that computationally
defines the relationship between underlying biological and
cognitive mechanisms to emerging behavior. Architectures,
such as ACT-R
          <xref ref-type="bibr" rid="ref4">(Anderson et al. 2004)</xref>
          and SOAR
          <xref ref-type="bibr" rid="ref24">(Laird,
Newell, and Rosenbloom 1987)</xref>
          , have long been a part of
HCI research to simulate users interacting with an interface.
For example, ACT-R models are used for usability testing
of menus
          <xref ref-type="bibr" rid="ref9">(Byrne 2001)</xref>
          , modeling how users detect
phishing websites (Williams and Li 2017), and detecting
situations with high cognitive load when using a smartphone
          <xref ref-type="bibr" rid="ref46 ref7">(Wirzberger and Russwinkel 2015)</xref>
          . Cognitive architectures
have been used with physiological data, such as eye tracking
information and fMRI, to map observed behavior the
underlying mental states and brain regions
          <xref ref-type="bibr" rid="ref41 ref45 ref46 ref7">(Tamborello and Byrne
2007; Borst and Anderson 2015)</xref>
          . Cognitive models,
combined with self-reported data from surveys and
physiological data, can provide a starting point for IML systems to
optimize their suggestions for the overall performance of a
human-machine team.
        </p>
        <p>These three different categories of cognitive feedback
– self reporting, implicit cognitive feedback, and modeled
cognitive feedback – delineate the possible ways in which
IML implementations may be centered around the analyst.
The categories are summarized in Table 1.</p>
        <p>
          Once cognitive feedback has been integrated into IML,
more conventional results such as classification accuracy
and overall corrections may be used to evaluate approaches
against their non-cognitive baseline. However, these
results may lack true insight into the purpose of the
humanmachine team. Measuring the cognitive load on human
subjects with more objective metrics of productivity would
provide more insight into the effectiveness of IML
implementations
          <xref ref-type="bibr" rid="ref1">(Alves et al. 2016)</xref>
          . Additionally, it is the analyst
themselves who must also evaluate the effectiveness of an IML
implementation, though this may take high levels of time
and effort
          <xref ref-type="bibr" rid="ref12 ref15">(Groce et al. 2013; Gillies et al. 2016)</xref>
          .
        </p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>A Future Driven by Cognitive Feedback</title>
      <p>We have presented a summary of interactive machine
learning along with several examples informing the state of the
art. After discussing the advantages of IML, the major
shortcomings and gaps were delineated. Finally, the implications
of cognitive feedback for IML implementations were
discussed to address the gaps. Though it may seem trivial to
study cognitive feedback as it relates to data science for
human-in-the-loop applications, there is a general lack of
such studies in the literature, especially for defense
applications. We hope this article will encourage research and
development in more IML for defense applications and more
research in how cognitive feedback may inform IML
implementations.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <surname>Alves</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Szpak</surname>
            ,
            <given-names>K. S.</given-names>
          </string-name>
          ; Gonc¸alves,
          <string-name>
            <given-names>J. L.</given-names>
            ;
            <surname>Sekino</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            ;
            <surname>Aquino</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ; e Castro, R. A.;
            <surname>Koglin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A</given-names>
            .;
            <surname>de Lima Fonseca</surname>
          </string-name>
          , N. B.; and
          <string-name>
            <surname>Mesa-Lao</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <year>2016</year>
          .
          <article-title>Investigating cognitive effort in post-editing: A relevance-theoretical approach</article-title>
          .
          <source>Eyetracking and Applied Linguistics</source>
          <volume>2</volume>
          :
          <fpage>109</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>Ananthanarayanan</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ; Bahl,
          <string-name>
            <given-names>P.</given-names>
            ; Bod´ık, P.;
            <surname>Chintalapudi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            ;
            <surname>Philipose</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ;
            <surname>Ravindranath</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            ; and
            <surname>Sinha</surname>
          </string-name>
          ,
          <string-name>
            <surname>S.</surname>
          </string-name>
          <year>2017</year>
          .
          <article-title>Realtime video analytics: The killer app for edge computing</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <source>computer 50</source>
          (10):
          <fpage>58</fpage>
          -
          <lpage>67</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <surname>Anderson</surname>
            ,
            <given-names>J. R.</given-names>
          </string-name>
          ; Bothell,
          <string-name>
            <surname>D.</surname>
          </string-name>
          ; Byrne,
          <string-name>
            <given-names>M. D.</given-names>
            ;
            <surname>Douglass</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            ;
            <surname>Lebiere</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          ; and Qin,
          <string-name>
            <surname>Y.</surname>
          </string-name>
          <year>2004</year>
          .
          <article-title>An integrated theory of the mind</article-title>
          .
          <source>Psychological review 111</source>
          <volume>(4)</volume>
          :
          <fpage>1036</fpage>
          -
          <lpage>1060</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <surname>Bachman</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Sordoni</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Trischler</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <year>2017</year>
          .
          <article-title>Learning algorithms for active learning</article-title>
          .
          <source>In Proceedings of the 34th International Conference on Machine Learning-</source>
          Volume
          <volume>70</volume>
          ,
          <fpage>301</fpage>
          -
          <lpage>310</lpage>
          . JMLR. org.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <surname>Blasch</surname>
            ,
            <given-names>E. P.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Breton</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ; Valin,
          <string-name>
            <given-names>P.</given-names>
            ; and
            <surname>Bosse</surname>
          </string-name>
          ,
          <string-name>
            <surname>E.</surname>
          </string-name>
          <year>2011</year>
          .
          <article-title>User information fusion decision making analysis with the c-ooda model</article-title>
          .
          <source>In 14th International Conference on Information Fusion</source>
          ,
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          . IEEE.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <surname>Borst</surname>
            ,
            <given-names>J. P.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Anderson</surname>
            ,
            <given-names>J. R.</given-names>
          </string-name>
          <year>2015</year>
          .
          <article-title>Using the ACT-R Cognitive Architecture in Combination With fMRI Data</article-title>
          . In Forstmann,
          <string-name>
            <given-names>B. U.</given-names>
            , and
            <surname>Wagenmakers</surname>
          </string-name>
          , E.-J., eds.,
          <article-title>An Introduction to Model-Based Cognitive Neuroscience</article-title>
          . New York, NY: Springer.
          <fpage>339</fpage>
          -
          <lpage>352</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <surname>Bottou</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <year>1998</year>
          .
          <article-title>Online learning and stochastic approximations</article-title>
          .
          <article-title>On-line learning in neural networks 17(9</article-title>
          ):
          <fpage>142</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <surname>Byrne</surname>
            ,
            <given-names>M. D.</given-names>
          </string-name>
          <year>2001</year>
          .
          <article-title>ACT-R/PM and menu selection: applying a cognitive architecture to HCI</article-title>
          .
          <source>International Journal of Human-Computer Studies</source>
          <volume>55</volume>
          (
          <issue>1</issue>
          ):
          <fpage>41</fpage>
          -
          <lpage>84</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <surname>Dudley</surname>
            ,
            <given-names>J. J.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Kristensson</surname>
            ,
            <given-names>P. O.</given-names>
          </string-name>
          <year>2018</year>
          .
          <article-title>A review of user interface design for interactive machine learning</article-title>
          .
          <source>ACM Trans. Interact. Intell. Syst</source>
          .
          <volume>8</volume>
          (
          <issue>2</issue>
          ):8:
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          :
          <fpage>37</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <surname>Fails</surname>
            ,
            <given-names>J. A.</given-names>
          </string-name>
          , and
          <string-name>
            <given-names>Olsen</given-names>
            <surname>Jr</surname>
          </string-name>
          ,
          <string-name>
            <surname>D. R.</surname>
          </string-name>
          <year>2003</year>
          .
          <article-title>Interactive machine learning</article-title>
          .
          <source>In Proceedings of the 8th international conference on Intelligent user interfaces</source>
          ,
          <fpage>39</fpage>
          -
          <lpage>45</lpage>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <surname>Gillies</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Fiebrink</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Tanaka</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Garcia</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Bevilacqua</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Heloir</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Nunnari</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Mackay</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ; Amershi,
          <string-name>
            <given-names>S.</given-names>
            ;
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <surname>B.</surname>
          </string-name>
          ; et al.
          <year>2016</year>
          .
          <article-title>Human-centred machine learning</article-title>
          .
          <source>In Proceedings of the 2016 CHI Conference Extended Abstracts on Human Factors in Computing Systems</source>
          ,
          <volume>3558</volume>
          -
          <fpage>3565</fpage>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <article-title>Gonza´lez-</article-title>
          <string-name>
            <surname>Rubio</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Casacuberta</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <year>2014</year>
          .
          <article-title>Costsensitive active learning for computer-assisted translation</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <source>Pattern Recognition Letters</source>
          <volume>37</volume>
          :
          <fpage>124</fpage>
          -
          <lpage>134</lpage>
          .
          <article-title>Partially Supervised Learning for Pattern Recognition</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <string-name>
            <surname>Groce</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Kulesza</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ; Zhang,
          <string-name>
            <given-names>C.</given-names>
            ;
            <surname>Shamasunder</surname>
          </string-name>
          ,
          <string-name>
            <surname>S.</surname>
          </string-name>
          ; Burnett,
          <string-name>
            <given-names>M.</given-names>
            ;
            <surname>Wong</surname>
          </string-name>
          , W.-K.;
          <string-name>
            <surname>Stumpf</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Das</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Shinsel</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Bice</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ; et al.
          <year>2013</year>
          .
          <article-title>You are the only possible oracle: Effective test selection for end users of interactive machine learning systems</article-title>
          .
          <source>IEEE Transactions on Software Engineering</source>
          <volume>40</volume>
          (
          <issue>3</issue>
          ):
          <fpage>307</fpage>
          -
          <lpage>323</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <surname>Gunning</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Aha</surname>
            ,
            <given-names>D. W.</given-names>
          </string-name>
          <year>2019</year>
          .
          <article-title>Darpa's explainable artificial intelligence program</article-title>
          .
          <source>AI</source>
          Magazine
          <volume>40</volume>
          (
          <issue>2</issue>
          ):
          <fpage>44</fpage>
          -
          <lpage>58</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <string-name>
            <surname>Hahnemann</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Beatty</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <year>1967</year>
          .
          <article-title>Pupillary responses in a pitch-discrimination task</article-title>
          .
          <source>Perception &amp; Psychophysics</source>
          <volume>2</volume>
          (
          <issue>3</issue>
          ):
          <fpage>101</fpage>
          -
          <lpage>105</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <string-name>
            <surname>Hart</surname>
            ,
            <given-names>S. G.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Staveland</surname>
            ,
            <given-names>L. E.</given-names>
          </string-name>
          <year>1988</year>
          .
          <article-title>Development of nasatlx (task load index): Results of empirical and theoretical research</article-title>
          . In Advances in psychology, volume
          <volume>52</volume>
          . Elsevier.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <string-name>
            <surname>Hess</surname>
            ,
            <given-names>E. H.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Polt</surname>
            ,
            <given-names>J. M.</given-names>
          </string-name>
          <year>1964</year>
          .
          <article-title>Pupil size in relation to mental activity during simple problem-solving</article-title>
          .
          <source>Science</source>
          <volume>143</volume>
          (
          <issue>3611</issue>
          ):
          <fpage>1190</fpage>
          -
          <lpage>1192</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <string-name>
            <surname>Hossain</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. D.</surname>
          </string-name>
          , and
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <year>2019</year>
          .
          <article-title>Segmentation for object-based image analysis (obia): A review of algorithms and challenges from remote sensing perspective</article-title>
          .
          <source>ISPRS Journal of Photogrammetry and Remote Sensing</source>
          <volume>150</volume>
          :
          <fpage>115</fpage>
          -
          <lpage>134</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          <string-name>
            <surname>Kabra</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Robie</surname>
            ,
            <given-names>A. A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Rivera-Alba</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Branson</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Branson</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <year>2013</year>
          .
          <article-title>Jaaba: interactive machine learning for automatic annotation of animal behavior</article-title>
          .
          <source>Nature methods 10</source>
          (1):
          <fpage>64</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          <string-name>
            <surname>Kahneman</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <year>1973</year>
          .
          <article-title>Attention and effort</article-title>
          , volume
          <volume>1063</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          <string-name>
            <surname>Koehn</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <year>2009</year>
          .
          <article-title>Statistical machine translation</article-title>
          . Cambridge University Press.
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          <string-name>
            <surname>Laird</surname>
            ,
            <given-names>J. E.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Newell</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ; and Rosenbloom,
          <string-name>
            <surname>P. S.</surname>
          </string-name>
          <year>1987</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          <string-name>
            <surname>SOAR</surname>
          </string-name>
          :
          <article-title>An architecture for general intelligence</article-title>
          .
          <source>Artificial Intelligence</source>
          <volume>33</volume>
          (
          <issue>1</issue>
          ):
          <fpage>1</fpage>
          -
          <lpage>64</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          <string-name>
            <surname>Langford</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Strehl</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <year>2007</year>
          .
          <article-title>Vowpal wabbit online learning project.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          <string-name>
            <surname>Lika</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Kolomvatsos</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Hadjiefthymiades</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          <article-title>Facing the cold start problem in recommender systems</article-title>
          .
          <source>Expert Syst. Appl</source>
          .
          <volume>41</volume>
          (
          <issue>4</issue>
          ):
          <fpage>2065</fpage>
          -
          <lpage>2073</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          <string-name>
            <surname>Lowenstein</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Feinberg</surname>
          </string-name>
          , R.; and
          <string-name>
            <surname>Loewenfeld</surname>
            ,
            <given-names>I. E.</given-names>
          </string-name>
          <year>1963</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          <article-title>Pupillary movements during acute and chronic fatigue: A new test for the objective evaluation of tiredness</article-title>
          .
          <source>Investigative Ophthalmology &amp; Visual Science</source>
          <volume>2</volume>
          (
          <issue>2</issue>
          ):
          <fpage>138</fpage>
          -
          <lpage>157</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          <string-name>
            <surname>Lu</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Shi</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Jia</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <year>2013</year>
          .
          <article-title>Online robust dictionary learning</article-title>
          .
          <source>In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition</source>
          ,
          <fpage>415</fpage>
          -
          <lpage>422</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          2019.
          <article-title>Towards an integrative theoretical framework of interactive machine learning systems</article-title>
          .
          <source>In Proceedings of the 27th European Conference on Information Systems(ECIS).</source>
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          <string-name>
            <surname>Michael</surname>
            , C. J.; Dennis,
            <given-names>S. M.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Maryan</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Irving</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ; and Palmsten,
          <string-name>
            <surname>M. L.</surname>
          </string-name>
          <year>2019</year>
          .
          <article-title>A general framework for humanmachine digitization of geographic regions from remotely sensed imagery</article-title>
          .
          <source>In Proceedings of the 27th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems</source>
          , SIGSPATIAL '
          <volume>19</volume>
          ,
          <fpage>259</fpage>
          -
          <lpage>268</lpage>
          . New York, NY, USA: ACM.
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          <string-name>
            <surname>Plonsky</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Erev</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ; Hazan,
          <string-name>
            <surname>T.</surname>
          </string-name>
          ; and Tennenholtz,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          <string-name>
            <surname>Quionero-Candela</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ; Sugiyama,
          <string-name>
            <given-names>M.</given-names>
            ;
            <surname>Schwaighofer</surname>
          </string-name>
          ,
          <string-name>
            <surname>A.</surname>
          </string-name>
          ; and Lawrence,
          <string-name>
            <surname>N. D.</surname>
          </string-name>
          <year>2009</year>
          .
          <article-title>Dataset shift in machine learning</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          <string-name>
            <surname>Raghavan</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Madani</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ; and Jones,
          <string-name>
            <surname>R.</surname>
          </string-name>
          <year>2006</year>
          .
          <article-title>Active learning with feedback on features and instances</article-title>
          .
          <source>Journal of Machine Learning Research</source>
          <volume>7</volume>
          (Aug):
          <fpage>1655</fpage>
          -
          <lpage>1686</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          <string-name>
            <surname>Robert</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ; Bu¨ttner, S.; Ro¨cker, C.; and
          <string-name>
            <surname>Holzinger</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          <article-title>Reasoning under uncertainty: Towards collaborative interactive machine learning</article-title>
          .
          <source>In Machine learning for health informatics</source>
          . Springer.
          <fpage>357</fpage>
          -
          <lpage>376</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          2012.
          <article-title>Combining psychological models with machine learning to better predict people's decisions</article-title>
          .
          <source>Synthese</source>
          <volume>189</volume>
          (
          <issue>1</issue>
          ):
          <fpage>81</fpage>
          -
          <lpage>93</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          <string-name>
            <surname>Talbot</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Kapoor</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Tan</surname>
            ,
            <given-names>D. S.</given-names>
          </string-name>
          <year>2009</year>
          .
          <article-title>Ensemblematrix: interactive visualization to support machine learning with multiple classifiers</article-title>
          .
          <source>In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems</source>
          ,
          <volume>1283</volume>
          -
          <fpage>1292</fpage>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref41">
        <mixed-citation>
          <string-name>
            <surname>Tamborello</surname>
            ,
            <given-names>F. P.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Byrne</surname>
            ,
            <given-names>M. D.</given-names>
          </string-name>
          <year>2007</year>
          .
          <article-title>Adaptive but nonoptimal visual search behavior with highlighted displays</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref42">
        <mixed-citation>
          <source>Cognitive Systems Research</source>
          <volume>8</volume>
          (
          <issue>3</issue>
          ):
          <fpage>182</fpage>
          -
          <lpage>191</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref43">
        <mixed-citation>
          <string-name>
            <surname>Teso</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Kersting</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <year>2019</year>
          .
          <article-title>Explanatory interactive machine learning</article-title>
          . In http://www. aies-conference.
        </mixed-citation>
      </ref>
      <ref id="ref44">
        <mixed-citation>
          <string-name>
            <surname>Turchetta</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Berkenkamp</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Krause</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <year>2019</year>
          .
          <article-title>Safe exploration for interactive machine learning</article-title>
          .
          <source>In Advances in Neural Information Processing Systems</source>
          ,
          <volume>2887</volume>
          -
          <fpage>2897</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref45">
        <mixed-citation>
          <string-name>
            <surname>Wilson</surname>
            ,
            <given-names>G. F.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Russell</surname>
            ,
            <given-names>C. A.</given-names>
          </string-name>
          <year>2007</year>
          .
          <article-title>Performance enhancement in an uninhabited air vehicle task using psychophysiologically determined adaptive aiding</article-title>
          .
          <source>Human factors 49</source>
          <volume>(6)</volume>
          :
          <fpage>1005</fpage>
          -
          <lpage>1018</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref46">
        <mixed-citation>
          <string-name>
            <surname>Wirzberger</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Russwinkel</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <year>2015</year>
          .
          <article-title>Modeling Interruption and Resumption in a Smartphone Task: An ACT-R Approach</article-title>
          .
          <source>Journal of Interactive Media</source>
          <volume>14</volume>
          (
          <issue>2</issue>
          ):
          <fpage>147</fpage>
          -
          <lpage>154</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref47">
        <mixed-citation>
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Weld</surname>
            ,
            <given-names>D. S.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Heer</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <year>2019</year>
          .
          <article-title>Local decision pitfalls in interactive machine learning: An investigation into feature selection in sentiment analysis</article-title>
          .
          <source>ACM Transactions on Computer-Human Interaction (TOCHI) 26</source>
          (
          <issue>4</issue>
          ):
          <fpage>24</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref48">
        <mixed-citation>
          <string-name>
            <surname>Zhang</surname>
            , R.; Liu,
            <given-names>Z.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Guan</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Zhang</surname>
          </string-name>
          , L.;
          <string-name>
            <surname>Hayhoe</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <article-title>M.;</article-title>
          and
          <string-name>
            <surname>Ballard</surname>
            ,
            <given-names>D. H.</given-names>
          </string-name>
          <year>2019</year>
          .
          <article-title>Atari-head: Atari human eye-tracking and demonstration dataset</article-title>
          . arXiv preprint arXiv:
          <year>1903</year>
          .06754.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>