<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Validating One-Class Active Learning with User Studies - a Prototype and Open Challenges</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Holger Trittenbach</string-name>
          <email>holger.trittenbach@kit.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Adrian Englhardt</string-name>
          <email>adrian.englhardt@kit.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Klemens B¨ohm</string-name>
          <email>klemens.boehm@kit.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Karlsruhe Institute of Technology (KIT)</institution>
          ,
          <addr-line>Karlsruhe</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <fpage>17</fpage>
      <lpage>31</lpage>
      <abstract>
        <p>Active learning with one-class classifiers involves users in the detection of outliers. The evaluation of one-class active learning typically relies on user feedback that is simulated, based on benchmark data. This is because validations with real users are elaborate. They require the design and implementation of an interactive learning system. But without such a validation, it is unclear whether the value proposition of active learning does materialize when it comes to an actual detection of outliers. User studies are necessary to find out when users can indeed provide feedback. In this article, we describe important characteristics and prerequisites of one-class active learning for outlier detection, and how they influence the design of interactive systems. We propose a reference architecture of a one-class active learning system. We then describe design alternatives regarding such a system and discuss conceptual and technical challenges. We conclude with a roadmap towards validating one-class active learning with user studies.</p>
      </abstract>
      <kwd-group>
        <kwd>Active learning</kwd>
        <kwd>One-class classification</kwd>
        <kwd>Outlier detection</kwd>
        <kwd>User study</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Active Learning is the paradigm to involve users in classifier training, to improve
classification results by means of feedback. An important application of active
learning is outlier detection. Here, one-class classifiers learn to discern inliers
from outliers. Examples are network security [
        <xref ref-type="bibr" rid="ref17 ref37">17, 37</xref>
        ] or fault monitoring [
        <xref ref-type="bibr" rid="ref45">45</xref>
        ]. In
these cases, one-class classifiers with active learning (OCAL) ask users to
provide a binary label (“inlier” or “outlier”) for some of the observations. Then they
use these labels in subsequent training iterations to learn an accurate decision
boundary. OCAL differs from other active learning applications such as balanced
binary or multi-class classification. This is because the strategies to select
observations for feedback (“query strategies”) take into account that outliers are
rare to non existent.
      </p>
      <p>
        Over the last years, several OCAL-specific query strategies have been
proposed. They focus on improving classification accuracy with minimal feedback [
        <xref ref-type="bibr" rid="ref15 ref16 ref18 ref3 ref45">3,
15, 16, 18, 45</xref>
        ]. To evaluate them experimentally, authors generally rely on data
sets with an available ground truth, in order to simulate user feedback. On the
c 2019 for this paper by its authors. Use permitted under CC BY 4.0.
V2alidatinTgriOttneneb-Caclhasest Aalc.tive Learning with User Studies
one hand, this seems to be convenient, since it allows to evaluate algorithmic
improvements without a user study. On the other hand, simulating user
feedback requires some assumptions on how users give feedback. These assumptions
are generally implicit, and authors do not elaborated on them, since they have
become widely accepted in the research community. In particular, there are two
fundamental assumptions behind active learning :
(Feedback) Users provide accurate feedback independently from the
presentation of the classification result and from the observation
selected for feedback (the “query”).
(Acceptance) Users do not question how the feedback provided changes
the classification results. Their motivation to provide
feedback, even for many observations, does not hinge on their
understanding of how their feedback influences the classifier.
      </p>
      <p>
        Think of an OCAL system that asks a user to provide a class label on a
20-dimensional real-valued vector, where features are the result of some
preprocessing, such as principal component analysis. We argue that, without further
information, users are unlikely to comprehend the query, and cannot provide
accurate feedback. This is in line with observations from literature [
        <xref ref-type="bibr" rid="ref14 ref33">14, 33</xref>
        ]. Even if
they could provide the feedback, any change in the classifier is not tangible. This
is because there is no suitable visualization or description of a 20-dim decision
boundary. We argue that users may question whether their feedback has a
positive effect on the classifier, or even any effect at all, lose interest, and eventually
discontinue to provide feedback.
      </p>
      <p>This is a peculiar situation: On the one hand, the value proposition of active
learning is to obtain helpful information from users that is not yet contained in
the training data. On the other hand, there currently is no validation whether one
can realize this value in an actual application. Such a validation would require
to implement an interactive OCAL system and to conduct a user study.</p>
      <p>
        However, such an experimental validation is difficult, since there are
several conceptual and technical issues. We have experienced this first hand, when
we have looked at smart meter data of an industrial production side [
        <xref ref-type="bibr" rid="ref42 ref7">7, 42</xref>
        ] to
identify unusual observations, and to collect respective ground truth labels from
human experts. In our preliminary experiments with this data, we found that
both active-learning assumptions do not hold in practice. In particular, we have
observed that domain experts ask for additional information such as
visualizations and explanations that go way beyond a simple presentation of classification
results. Since there is an over-reliance on active-learning assumptions, only little
effort has been spent on making OCAL interpretable, comprehensible, and
usable. So it is unclear what the minimal requirements behind an OCAL system
are to carry out a user study. Second, there are conceptual issues that are in
the way of implementing OCAL systems. One issue is that the design space of
OCAL systems is huge. It requires to define a learning scenario, to choose a
suitable classifier and a learning strategy, as well as selecting multiple
hyperparameter values. In addition, there may be several conflicting objectives: One may
strive to improve classification accuracy. Another objective may be to use OCAL
      </p>
    </sec>
    <sec id="sec-2">
      <title>Backend</title>
      <p>One-Class Classi ers
Query Strategies
Explanations
l
j
.</p>
      <p>I
P
A
l
a
c</p>
      <p>O</p>
      <p>Data
Experiments</p>
    </sec>
    <sec id="sec-3">
      <title>Frontend</title>
      <p>Operator Interface
Participant Interface</p>
      <p>Operator
The purpose of an OCAL system is to facilitate experiments with several users.
An experiment is a specific technical configuration, i.e., a data set, a classifier, a
query strategy, and one or more users, the participants, who provide feedback.</p>
      <p>An OCAL system consists of several modules. Participants interact with
the system through a participant interface that visualizes information on active
V4alidatinTgriOttneneb-Caclhasest Aalc.tive Learning with User Studies
learning iterations, such as the classification result and the progress of the
experiment. The training of the classifier, query selection, and the preparation of
additional information such as visualizations and explanations take place in an
algorithm backend. Finally, there is a human operator who configures, monitors
and evaluates the experiments through an operator interface. This typically is
the researcher who conducts the experiments. Figure 1 is an overview of the
system architecture. In the following, we describe the different modules and link
them to our prototype implementation.</p>
      <p>Algorithm Backend: On a technical level, the algorithm backend consists of
a classifier module SVDD.jl 1 and a module OneClassActiveLearning.jl 2, which
implements active learning components such as the query strategies. A third
module provides additional information, e.g., classifier visualizations. For our
prototype, we have implemented the classifiers, query strategies and basic
visualization information in OcalAPI.jl 3, a ready-to-use JSON REST API. This
decoupling allows to re-use the algorithm backend independent of the participant
and operator interface.</p>
      <p>Operator Interface: The operator interface allows an operator to configure
so-called experiment setups. A setup consists of a data set, a parameterized
classifier and a query strategy. Depending on the research question, the operator
may also configure which information is displayed in the participant interface.
This gives way to A/B tests, to, say, validate if a certain visualization has an
effect on feedback quality. The operator can invite several users to participate in
an experiment run, i.e., an instantiation of an experiment setup. He can monitor
and inspect the experiment runs in an overview panel and export experiment
data for further analysis.</p>
      <p>Participant Interface: The participant interface has two functions. First, it
is an input device to collect feedback during the experiment. Second, it provides
the participants with information that supports them to provide educated
feedback. For instance, this may be a visualization of a classifier, a view on the raw
data or a history of classification accuracy over the past iterations. The
participant then provides feedback for some observations. During this process, the
interface captures user interactions, e.g., mouse movement and selection. When
the query budget or time limit is not exhausted, the participant proceeds with
the next iteration.</p>
      <p>Our implementation of the interfaces is preliminary, since there are several
open challenges, both conceptual and technical (see Section 3). We plan to make
1 https://github.com/englhardt/SVDD.jl
2 https://github.com/englhardt/OneClassActiveLearning.jl
3 https://github.com/englhardt/OcalAPI.jl
it publicly available in the future as well. An important takeaway from this
section is an intuition about how OCAL systems can be designed, on an
architectural level. This intuition may be useful to understand the following discussions
on the design space of OCAL systems and the challenges related to the three
modules.
3</p>
      <sec id="sec-3-1">
        <title>Design Decisions for OCAL Systems</title>
        <p>The design and implementation of OCAL systems are inherently
interdisciplinary and require expertise from several areas, including data mining,
humancomputer interaction, UX-design, and knowledge of the application domain.
Although all disciplines are important, we now focus on the data mining
perspective. We first discuss different types of interaction and elaborate on the design
options for one-class classifiers and query strategies. We then present different
options to prepare information for users during the learning iterations. Finally,
we elaborate on several technical challenges when implementing OCAL systems.
3.1</p>
        <sec id="sec-3-1-1">
          <title>Type of Interaction</title>
          <p>
            The common definition of active learning is that a query strategy selects one or
more observations for feedback. So, strictly speaking, a user does not have the
option to also give feedback on other observations not selected by the system.
However, there are related disciplines that do away with this restriction. For
instance, one research direction is Visual Interactive Analytics (VIA) [
            <xref ref-type="bibr" rid="ref25 ref43 ref8">8, 25, 43</xref>
            ],
where a user interactively explores outliers in a data set. VIA systems provide
different kinds of visualization to assist users in identifying outliers, in particular
with high-dimensional data sets. The unification of active learning and VIA is
Visual Inter-Active Labeling (VIAL) [
            <xref ref-type="bibr" rid="ref27 ref6">6,27</xref>
            ]. VIAL combines active learning with
user-supporting visualizations from the VIA community. Variants of VIAL and
active learning are conceivable as well. For instance, instead of asking for labels
of specific observations, the query strategy could provide a set of observations
from which users can select one or more to label.
          </p>
          <p>
            It is an open question in which cases one should use VIAL or active learning.
A user study in [
            <xref ref-type="bibr" rid="ref5">5</xref>
            ] indicates that users label more observations if they are free
to choose the observations. However, the resulting classifier accuracy is higher
with an AL query strategy. It is unclear whether these insights transfer to
outlier detection where classes are unbalanced. In fact, we see this as one of the
overarching questions to answer with user studies.
3.2
          </p>
        </sec>
        <sec id="sec-3-1-2">
          <title>Type of Feedback</title>
          <p>
            In this article, feedback is binary, i.e., users decide whether an observation
belongs to the inlier or outlier class. However, other types of feedback are
conceivable as well. For instance, in multi-class settings, the system may ask users to
state to which classes an observation does not belong [
            <xref ref-type="bibr" rid="ref10">10</xref>
            ]. Another example is to
V6alidatinTgriOttneneb-Caclhasest Aalc.tive Learning with User Studies
ask users for feedback on features, as opposed to instances [
            <xref ref-type="bibr" rid="ref13">13</xref>
            ]. Existing OCAL
approaches in turn focus on binary feedback. It is an open question if and how
OCAL can benefit from allowing for different types of feedback.
3.3
          </p>
        </sec>
        <sec id="sec-3-1-3">
          <title>OCAL Design Space</title>
          <p>
            An OCAL system consists of three building blocks: the learning scenario, the
classifier, and the query strategy. In brief, a learning scenario are underlying
assumptions about the application and user interaction. This includes the
feedback type, e.g., sequential labels, the budget available for feedback, as well as
assumptions on the data distribution, the objective of the learning process, e.g.,
to improve the accuracy of the classifier, and an initial setup, which includes how
many labels are available when the active learning starts. In addition, there are
several semi-supervised classifiers, such as SVDDneg [
            <xref ref-type="bibr" rid="ref39">39</xref>
            ], and query strategies,
e.g., high-confidence sampling [
            <xref ref-type="bibr" rid="ref3">3</xref>
            ], which one can combine almost arbitrarily with
any of the learning scenarios. Almost all classifiers and query strategies require
to set additional hyperparameters. Their value can have significant influence on
result quality, and a poor choice may bog it down. Moreover, a good query
strategy and hyperparemeter values may also depend on the active learning progress,
i.e., the number of labels already provided by the user.
          </p>
          <p>
            Navigating this design space is challenging, and it is generally not feasible to
consider and evaluate all possible combinations. Although there is an overview
and a benchmark on OCAL [
            <xref ref-type="bibr" rid="ref41">41</xref>
            ], a good solution still is application-specific and
may require fine-tuning of several components.
3.4
          </p>
        </sec>
        <sec id="sec-3-1-4">
          <title>Preparation of Information</title>
          <p>Classifier training and query selection produce a lot of data. On a fine-granular
level, this includes the parameterized decision function for the classifier and
informativeness scores for the query strategy. After processing this data, query
strategies select the most informative instances and predict a label for each
observation. In general, this data can be processed and enriched in many ways
before presenting it to a user. On a coarse level, one can provide users with
additional information, such as explanations of the classifier or contextual
information on the learning progress. We now discuss several types of information
to present during an active learning iteration: the query, the result, black-box
explanations and contextual information.</p>
          <p>Query presentation: After selecting observations for feedback, “queries” in
short, they must be presented to the user. In general, there are two
representations of a query. First, the query has a raw-data representation. Examples are
text documents, multimedia files, multi-dimensional time series of real-valued
sensors, or sub-graphs of a network. Second, the data often is pre-processed to
a feature representation, a real-valued vector that the classifier can process. In
principle, queries can be presented to users in either representation. Our
experience is that domain experts are more familiar with raw data and demand it
even if the feature representation is interpretable.</p>
          <p>
            Next, one can provide context information for queries. For an individual
instance, one can show the nearest neighbors of the query or a difference to
prototypes of both classes. Another approach is to use visualization techniques
for high-dimensional data [
            <xref ref-type="bibr" rid="ref28 ref32">28, 32</xref>
            ] to highlight the query. One can also visualize
the score distribution over all candidate queries. Depending on the type of the
query strategy, it also is possible to generate heatmaps that indicate areas in the
data space with high informativeness [
            <xref ref-type="bibr" rid="ref44">44</xref>
            ] together with the query.
Result presentation: The presentation of a classification result largely
depends on the classifier. With OCAL, the classifiers predominantly used rely on
the notion of Support Vector Data Description (SVDD) [
            <xref ref-type="bibr" rid="ref39">39</xref>
            ]. In a nutshell, SVDD
is an optimization problem to fit a hypersphere around the data, while allowing
a small percentage of the data, the outliers, to lie outside the hypersphere. By
using the kernel trick, the decision boundaries can become of arbitrary shape.
So a natural presentation of SVDD is a contour plot that shows distances to
the decision boundary. However, when data has more than two dimensions,
contour plots are not straightforward. The reason is that contour plots rely on the
distance to the decision boundary for a two-dimensional grid of observations
(x1, x2). However, the distance depends on the full vector (x1, x2, . . . , xn) and
thus cannot be computed for low-dimensional projections. One remedy would be
to train a classifier for each of the projections to visualize. However, the classifier
trained on the projection may differ significantly from the classifier trained on
all dimensions. So a two-dimensional contour plot may have very little benefit.
With common implementations of one-class classifiers, one is currently restricted
to present results as plain numeric values, raw data, and predicted labels.
Black-Box Explanations: Orthogonal to inspecting the queries and the
classification result, there are several approaches to provide additional explanations
of the classification result. The idea is to treat the classifier, or more generally
any predictive model, as a black box, and generate post-hoc explanations for the
prediction of individual observations. This is also called local explanation, since
explanations differ between instances. Recently, CAIPI, a local explainer based
on the popular explanation framework LIME [
            <xref ref-type="bibr" rid="ref30">30</xref>
            ], has been proposed to explain
classification results in an active learning setting [
            <xref ref-type="bibr" rid="ref40">40</xref>
            ]. The idea behind CAIPI is
to provide the user with explanations for the prediction of a query and ask them
to correct wrong explanations. Another application of LIME is to explain why an
observation has been selected as a query [
            <xref ref-type="bibr" rid="ref29">29</xref>
            ]. The idea behind this approach is to
explain the informativeness of a query by its neighborhood. The authors use
uncertainty sampling, and this approach may also work with other query strategies,
such as high-confidence sampling [
            <xref ref-type="bibr" rid="ref3">3</xref>
            ]. However, with more complex query
strategies, for instance ones that incorporate local neighborhoods [
            <xref ref-type="bibr" rid="ref45">45</xref>
            ] or probability
densities [
            <xref ref-type="bibr" rid="ref16">16</xref>
            ], applying LIME may not be straightforward. For outlier detection,
V8alidatinTgriOttneneb-Caclhasest Aalc.tive Learning with User Studies
there exist further, more specific approaches to generate explanations for
outlierness. An example is to visualize two-dimensional projections for input features
that contribute most to an outlier score [
            <xref ref-type="bibr" rid="ref19">19</xref>
            ]. Other examples are methods from
the VIA community that allow users to explore outliers interactively [
            <xref ref-type="bibr" rid="ref25 ref43 ref8">8, 25, 43</xref>
            ].
Contextual Information: The participant interface can also provide
additional information that spans several active learning iterations. For instance, the
interface can give users access to the classification history, allow them to revisit
their previous responses, and give them access to responses of other users, if
available. This can entail several issues, such as how to combine possibly
diverging responses from different users, and the question whether users will be
biased by giving them access to feedback of others. Studying such issues is
focus of collaborative interactive learning [
            <xref ref-type="bibr" rid="ref9">9</xref>
            ]. Others have proposed to give users
access to 2D scatter plots of the data, the confusion matrix and the progress of
classification accuracy on labeled data [
            <xref ref-type="bibr" rid="ref24">24</xref>
            ]. In this case, accuracy measures may
be biased. For instance, after collecting a ground truth for the first few labels,
accuracy may be very high. It may decrease when more labels become available,
and the labeled sample covers a larger share of the data space. So it remains
an open question whether contextual information will indeed support users to
provide accurate feedback.
          </p>
          <p>
            To conclude, one faces many options in the design of OCAL systems. In
particular, there are many approaches to support users with information so that
they can make informed decisions on the class label. However, the approaches
discussed have not yet been evaluated by means of user studies. Instead, they
are limited to a theoretical discussion, simulated feedback based on benchmark
data, or pen and paper surveys [
            <xref ref-type="bibr" rid="ref40">40</xref>
            ]. It is largely unclear which methods do
enable users to provide feedback and indeed improve the feedback collected.
3.5
          </p>
        </sec>
        <sec id="sec-3-1-5">
          <title>Technical Challenges</title>
          <p>
            Active learning induces several technical requirements to make systems
interactive, and to collect user feedback. Most requirements are general for active
learning systems But their realization with one-class classifiers is difficult.
Cold Start In most cases, active learning starts with a fully unsupervised
setting, i.e., there is no labeled data available. This restricts the possible
combinations of classifiers and query strategies in two cases. First, some query strategies,
e.g., sampling close to the decision boundary, require a trained one-class
classifier to calculate informativeness. In this case, the classifier must be applicable
both in an unsupervised and a supervised setting. Second, some query strategies
rely on labeled data, e.g., when estimating probability densities for the inlier
class [
            <xref ref-type="bibr" rid="ref15 ref16">15, 16</xref>
            ]. In this case, one cannot calculate informativeness without labels.
Current benchmarks mostly avoid this issue by simply assuming that some
observations from each class are already labeled. In a real system, one must think
about how to obtain the initially labeled observations [
            <xref ref-type="bibr" rid="ref2 ref21">2, 21</xref>
            ]. One option would
be to start with a query strategy that does not require any label, such as random
sampling, and switch to a more sophisticated strategy once there are sufficiently
many labels. Another option is to let users pick the observations to label in
the beginning, and then switch to an active learning strategy [
            <xref ref-type="bibr" rid="ref2 ref6">2, 6</xref>
            ]. However,
deciding when to do switches between query strategies with OCAL is an open
question.
          </p>
          <p>
            Batch Query Selection Currently, query selection for one-class classifiers is
sequential, i.e., for one observation at a time. However, this sequentiality may
have several disadvantages, such as frequent updating and re-training of the
oneclass classifier. Further, it might be easier for users to label several observations in
a batch than one observation at a time [
            <xref ref-type="bibr" rid="ref34">34</xref>
            ]. This may be the case when showing a
diverse set of observations helps a user to develop an intuition regarding the data
set. However, there currently are no strategies to select multiple observations in
batches with one-class classifiers. An open question is whether strategies that
have been proposed for other use cases, such as multi-class classification [
            <xref ref-type="bibr" rid="ref12">12</xref>
            ], are
applicable with one-class classifiers.
          </p>
          <p>Incremental Learning The runtime for updating a classifier constrains the
frequency of querying the user. In particular, excessive runtimes for classifier
training result in long waiting times and do away with interactivity. Intuitively,
there is an upper limit that users are willing to wait, but the specific limit
depends on the application.</p>
          <p>
            Several strategies are conceivable to mitigate runtime issues. First, one can
rely on incremental learning algorithms [
            <xref ref-type="bibr" rid="ref20">20</xref>
            ]. However, state-of-the-art one-class
classifiers like SSAD [
            <xref ref-type="bibr" rid="ref18">18</xref>
            ] have been proposed without any feature for
incremental learning. Second, one can sub-sample to reduce the number of training
observations. Several strategies have been proposed explicitly for one-class
classifiers [
            <xref ref-type="bibr" rid="ref23 ref26 ref38">23, 26, 38</xref>
            ]. But to our knowledge, there are no studies that combine
subsampling with OCAL. Finally, one can use speculative execution to pre-compute
the classifier update for both outcomes (inlier or outlier) while the user is
deciding on a label [
            <xref ref-type="bibr" rid="ref36">36</xref>
            ]. While such a strategy requires additional computational
resources, it might reduce waiting times significantly and improve interactivity.
The open question is how to proceed with pre-computing when the look-ahead
l is more than one feedback iteration. This is a combinatorial problem, and
precomputing all 2l learning paths is intractable. Instead, one may use conditional
probabilities to pre-compute only the most likely search paths. However, there
currently is no method to plan pre-computation beyond l = 1. If users select
observations to label by themselves, pre-computation would require to compute
classifier update for all observations and outcomes, which is infeasible. Thus,
there is a trade-off between giving users flexibility to decide freely on which
observations to label, and the capabilities of pre-computation.
V1a0lidatinTgriOttneneb-Caclhasest Aalc.tive Learning with User Studies
          </p>
          <p>
            Evaluation at Runtime Without a good quality estimate, it is impossible to
know whether the feedback obtained from a user already is sufficient [
            <xref ref-type="bibr" rid="ref2">2</xref>
            ], i.e., the
one-classifier has converged, and additional feedback would not alter the decision
boundary any further. However, evaluating the classification quality of OCAL
at runtime is difficult [
            <xref ref-type="bibr" rid="ref22">22</xref>
            ]. This issue exists in both, when benchmarking with
simulated feedback, and in real systems – here, we focus on the latter. Users
may become frustrated if they face periods where their feedback does not have
any effect.
          </p>
          <p>However, showing users any estimated classification quality is difficult for two
reasons. First, there might be a short term bias, i.e., the classifier performance
might fluctuate significantly. This may be irritating, and it may be difficult to
assess for the user. Second, the number of observations in the ground truth
increases over time. With only a few labeled observations, the quality estimates
may have a large error. This error may reduce with more iterations. So the open
question is how to estimate classification quality reliably, and how to adapt
these quality estimates during learning. One conceivable option is to switch
between exploration and exploitation, i.e., switch from querying for examples
that improve classification quality to selection strategies that improve the quality
estimate of the classifier. However, there currently is no such switching method
for OCAL.</p>
          <p>Management of Data Flows Developing an active learning system also
requires a sound software architecture. Although this is not a research challenge
per se, there are several aspects to consider when implementing OCAL systems.
One key aspect is the management of data flows. In particular, with a distributed
application, see Section 2, there are several locations where one has to retain the
data set, the classifier, the predictions, and the informativeness scores. For large
data sets in particular, transferring data between a client and a backend or
loading data sets from disc may affect runtimes significantly. This calls for efficient
data caching. Further, one must decide where computations take place. For
instance, to visualize contour plots, one must predict the decision boundary for a
grid of observations, possibly in multiple projections of the data. In this case,
transferring the model over the network may be very little overhead. This can be
an efficient strategy when evaluating the model for an observation is cheap. This
is the case with SVDD, since the model consists of only a few support vectors.
With multi-user studies, one may even reuse trained classifiers and
informativeness scores from other user sessions with an equivalent feedback history. In this
case, it might be more efficient to pre-compute grid predictions in the backend.
So there are several trade-offs and factors that determine an efficient data flow.
There currently is no overview on these trade-offs. It also is unclear how they
affect design decisions for OCAL systems.</p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>Validating OCAL with User Studies</title>
        <p>
          There are a few active learning user studies which have been conducted for special
use cases, such as text corpus annotation [
          <xref ref-type="bibr" rid="ref1 ref11 ref31">1, 11, 31</xref>
          ] and network security [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ].
However, it is unclear how findings relate to outlier detection with OCAL –
the previous sections illustrate the peculiarities of this application. Further, the
plethora of design options make user studies with OCAL systems particularly
challenging.
        </p>
        <p>
          Addressing all of the design options at once is not feasible, since there are
too many combinations of classifiers, query strategies and ways to prepare
information for users. So we propose to start with a narrow use case and to increase
the complexity of the OCAL system step-wise. Specifically, we have identified
the following steps towards a validation of OCAL in real applications.
(i) Simplified Use Case: Much of the value of active learning is in domains
where obtaining labels is difficult, even for domain experts. However, we
argue that one should identify a use case that many people can easily
relate to. This has several advantages. First, we deem reproducibility
more important than to obtain sophisticated insights on very special use
cases. User studies are easier to reproduce when they do not depend
on specific domain expertise. Further, when relationships in data are well
understood, one can more easily judge whether the presentation of queries
and results is accurate. So we argue to base a validation of OCAL on
standard benchmark data, for instance the hand-written digit image data
set MNIST4. Such a simplification also includes to fix the details of the
feedback process, for instance to “sequential feedback” and “no initial
labels”. If necessary, one should downsample data sets so that runtimes
of classifiers and query strategies are not a bottleneck.
(ii) Validation of Information Presented: The next step is to identify
situations when users can give accurate feedback. Since the focus is to
validate a learning system with users, one should start with a data set with
available ground truth and select the best combination of classifier and
query strategy in an experimental benchmark. This might seem
counterintuitive at first sight. In a real application, there generally are not
sufficiently many labels available to conduct such a benchmark – in fact, this
may even be the motivation for active learning in the first place [
          <xref ref-type="bibr" rid="ref2 ref35">2, 35</xref>
          ].
However, we argue that this is a necessary step to break the mutual
dependency between selecting a good setup and collecting labels. Given a
combination of classifier and query strategy, one can then apply different
query and result presentations and work with explanations and
contextual information. By evaluating this step with user experiments, one can
derive assumptions which, if met, enable users to provide accurate
feedback.
(iii) Validation of Classifier and Learning Strategy: Based on these
assumptions, one can vary the dimensions that have been fixed beforehand. This
4 http://yann.lecun.com/exdb/mnist/
V1a2lidatinTgriOttneneb-Caclhasest Aalc.tive Learning with User Studies
is, one fixes the information presented to the user and varies the query
strategies and classifiers. Further, one may validate specific extensions
such as batch query strategies.
(iv) Generalization: The first step of generalization is to scale the experiments
to a large number of observations, using the techniques discussed in
Section 3.5. Finally, one can then validate the approach on similar data sets,
e.g., on different image data.
        </p>
        <p>We expect the findings from these steps to be two-fold. On the one hand, we
expect insights that are independent from the use case. For instance, whether
scalability techniques are useful is likely to be use-case independent. On the other
hand, many findings may depend on the type of data at hand. Explanations based
on image data may be very different from the ones for, say, time series data.</p>
        <p>Our OCAL system prototype already includes different classifiers and query
strategies, see Section 2. So, in general, any researcher can already use our system
to conduct Step (i) and the pre-selection of the query strategy and classifier
information required for Step (ii). Regarding our prototype, the next steps are
to select and implement a working set of query and result presentations, as well
as to include black-box explainers and contextual information.
5</p>
      </sec>
      <sec id="sec-3-3">
        <title>Conclusions</title>
        <p>Validating One-Class Active Learning through user studies is challenging. One
reason is that there are several open conceptual and technical challenges in
the design and implementation of interactive learning systems. This article has
featured a systematic overview of these challenges, and we have pointed out
open research questions with one-class active learning. Next, we have sketched
an architecture of a one-class active learning system, which we have implemented
as a prototype. Based on it, we propose a roadmap towards validating one-class
active learning with user studies.</p>
        <p>Acknowledgement This work was supported by the German Research
Foundation (DFG) as part of the Research Training Group GRK 2153: Energy Status
Data – Informatics Methods for its Collection, Analysis and Exploitation.
V1a4lidatinTgriOttneneb-Caclhasest Aalc.tive Learning with User Studies</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Arora</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nyberg</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          , Ros´e,
          <string-name>
            <surname>C.P.</surname>
          </string-name>
          :
          <article-title>Estimating annotation cost for active learning in a multi-annotator environment</article-title>
          .
          <source>In: NAACL Workshop</source>
          . pp.
          <fpage>18</fpage>
          -
          <lpage>26</lpage>
          . ACL (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Attenberg</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Provost</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>Inactive learning?: Difficulties employing active learning in practice</article-title>
          .
          <source>SIGKDD Explor. Newsl</source>
          .
          <volume>12</volume>
          (
          <issue>2</issue>
          ),
          <fpage>36</fpage>
          -
          <lpage>41</lpage>
          (
          <year>2011</year>
          ). https://doi.org/10.1145/1964897.1964906
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Barnab</surname>
          </string-name>
          <article-title>´e-</article-title>
          <string-name>
            <surname>Lortie</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bellinger</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Japkowicz</surname>
          </string-name>
          , N.:
          <article-title>Active learning for One-Class classification</article-title>
          .
          <source>In: ICMLA</source>
          . pp.
          <fpage>390</fpage>
          -
          <lpage>395</lpage>
          . IEEE (
          <year>2015</year>
          ). https://doi.org/10.1109/ICMLA.
          <year>2015</year>
          .167
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Beaugnon</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chifflier</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bach</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>End-to-end active learning for computer security experts</article-title>
          .
          <source>In: AAAI Workshop</source>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Bernard</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hutter</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zeppelzauer</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fellner</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sedlmair</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Comparing visual-interactive labeling with active learning: An experimental study</article-title>
          .
          <source>Trans. Vis. Comput. Graph</source>
          .
          <volume>24</volume>
          (
          <issue>1</issue>
          ),
          <fpage>298</fpage>
          -
          <lpage>308</lpage>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Bernard</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zeppelzauer</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sedlmair</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Aigner</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          :
          <article-title>Vial: a unified process for visual interactive labeling</article-title>
          .
          <source>The Visual Computer</source>
          <volume>34</volume>
          (
          <issue>9</issue>
          ),
          <fpage>1189</fpage>
          -
          <lpage>1207</lpage>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Bischof</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Trittenbach</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vollmer</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Werle</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Blank</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          , Bo¨hm, K.:
          <article-title>Hipe: An energy-status-data set from industrial production</article-title>
          .
          <source>In: Proceedings of the Ninth International Conference on Future Energy Systems</source>
          . pp.
          <fpage>599</fpage>
          -
          <lpage>603</lpage>
          . ACM (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8. Bo¨gl,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Filzmoser</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Gschwandtner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            ,
            <surname>Lammarsch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            ,
            <surname>Leite</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.A.</given-names>
            ,
            <surname>Miksch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            ,
            <surname>Rind</surname>
          </string-name>
          ,
          <string-name>
            <surname>A.</surname>
          </string-name>
          :
          <article-title>Cycle plot revisited: Multivariate outlier detection using a distance-based abstraction</article-title>
          .
          <source>Computer Graphics Forum</source>
          <volume>36</volume>
          (
          <issue>3</issue>
          ),
          <fpage>227</fpage>
          -
          <lpage>238</lpage>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Calma</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Leimeister</surname>
            ,
            <given-names>J.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lukowicz</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Oeste-Reiss</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Reitmaier</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schmidt</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sick</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stumme</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zweig</surname>
            ,
            <given-names>K.A.</given-names>
          </string-name>
          :
          <article-title>From active learning to dedicated collaborative interactive learning</article-title>
          .
          <source>In: ARCS 2016; 29th International Conference on Architecture of Computing Systems</source>
          . pp.
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          (
          <year>Apr 2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Cebron</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Richter</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lienhart</surname>
          </string-name>
          , R.:
          <article-title>“I can tell you what it's not”: active learning from counterexamples</article-title>
          .
          <source>Prog. Artif. Intell</source>
          .
          <volume>1</volume>
          (
          <issue>4</issue>
          ),
          <fpage>291</fpage>
          -
          <lpage>301</lpage>
          (
          <year>2012</year>
          ). https://doi.org/10.1007/s13748-012-0023-9
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Choi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Park</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Choo</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hong</surname>
            ,
            <given-names>S.R.</given-names>
          </string-name>
          : Aila:
          <article-title>Attentive interactive labeling assistant for document classification through attention-based deep neural networks</article-title>
          .
          <source>In: CHI. ACM</source>
          (
          <year>2019</year>
          ). https://doi.org/10.1145/3290605.3300460
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Demir</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Persello</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bruzzone</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Batch-Mode Active-Learning methods for the interactive classification of remote sensing images</article-title>
          .
          <source>Trans. Geosci. Remote Sens</source>
          .
          <volume>49</volume>
          (
          <issue>3</issue>
          ),
          <fpage>1014</fpage>
          -
          <lpage>1031</lpage>
          (
          <year>2011</year>
          ). https://doi.org/10.1109/TGRS.
          <year>2010</year>
          .2072929
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Druck</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Settles</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>McCallum</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Active learning by labeling features</article-title>
          .
          <source>In: EMNLP</source>
          . pp.
          <fpage>81</fpage>
          -
          <lpage>90</lpage>
          . ACL (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Endert</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hossain</surname>
            ,
            <given-names>M.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ramakrishnan</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>North</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fiaux</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Andrews</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>The human is the loop: new directions for visual analytics</article-title>
          .
          <source>J. Intell. Inf. Syst</source>
          .
          <volume>43</volume>
          (
          <issue>3</issue>
          ),
          <fpage>411</fpage>
          -
          <lpage>435</lpage>
          (
          <year>2014</year>
          ). https://doi.org/10.1007/s10844-014-0304-9
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Ghasemi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manzuri</surname>
            ,
            <given-names>M.T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rabiee</surname>
            ,
            <given-names>H.R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rohban</surname>
            ,
            <given-names>M.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Haghiri</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Active oneclass learning by kernel density estimation</article-title>
          .
          <source>In: MLSP Workshop</source>
          . pp.
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          (
          <year>2011</year>
          ). https://doi.org/10.1109/MLSP.
          <year>2011</year>
          .6064627
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Ghasemi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rabiee</surname>
            ,
            <given-names>H.R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fadaee</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manzuri</surname>
            ,
            <given-names>M.T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rohban</surname>
            ,
            <given-names>M.H.</given-names>
          </string-name>
          :
          <article-title>Active learning from positive and unlabeled data</article-title>
          .
          <source>In: ICDM Workshop</source>
          . pp.
          <fpage>244</fpage>
          -
          <lpage>250</lpage>
          (
          <year>2011</year>
          ). https://doi.org/10.1109/ICDMW.
          <year>2011</year>
          .20
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17. G¨ornitz, N.,
          <string-name>
            <surname>Kloft</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rieck</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brefeld</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          :
          <article-title>Active learning for network intrusion detection</article-title>
          .
          <source>In: AiSec Workshop</source>
          . ACM (
          <year>2009</year>
          ). https://doi.org/10.1145/1654988.1655002
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18. G¨ornitz, N.,
          <string-name>
            <surname>Kloft</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rieck</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brefeld</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          :
          <article-title>Toward supervised anomaly detection</article-title>
          .
          <source>J. Artif. Intell. Res</source>
          .
          <volume>46</volume>
          ,
          <fpage>235</fpage>
          -
          <lpage>262</lpage>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Gupta</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Eswaran</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shah</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Akoglu</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Faloutsos</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Beyond outlier detection: LOOKOUT for pictorial explanation</article-title>
          .
          <source>In: ECML</source>
          . pp.
          <fpage>122</fpage>
          -
          <lpage>138</lpage>
          . Springer (
          <year>2018</year>
          ). https://doi.org/10.1145/1235
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Kefi-Fatteh</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ksantini</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          , Kaaˆniche, M.B.,
          <string-name>
            <surname>Bouhoula</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>A novel incremental one-class support vector machine based on low variance direction</article-title>
          .
          <source>Pattern Recognit</source>
          .
          <volume>91</volume>
          ,
          <fpage>308</fpage>
          -
          <lpage>321</lpage>
          (
          <year>2019</year>
          ). https://doi.org/10.1016/j.patcog.
          <year>2019</year>
          .
          <volume>02</volume>
          .027
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Kottke</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Calma</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huseljic</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Krempl</surname>
            ,
            <given-names>G.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sick</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Challenges of reliable, realistic and comparable active learning evaluation</article-title>
          .
          <source>In: Proceedings of the Workshop and Tutorial on Interactive Adaptive Learning</source>
          . pp.
          <fpage>2</fpage>
          -
          <lpage>14</lpage>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Kottke</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schellinger</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huseljic</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sick</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Limitations of assessing active learning performance at runtime</article-title>
          . arXiv:
          <year>1901</year>
          .
          <volume>10338</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Krawczyk</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Triguero</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          , Garc´ıa,
          <string-name>
            <surname>S.</surname>
          </string-name>
          , Wo´zniak,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Herrera</surname>
          </string-name>
          ,
          <string-name>
            <surname>F.</surname>
          </string-name>
          :
          <article-title>Instance reduction for one-class classification</article-title>
          .
          <source>Knowl. Inf. Syst</source>
          .
          <volume>59</volume>
          (
          <issue>3</issue>
          ),
          <fpage>601</fpage>
          -
          <lpage>628</lpage>
          (
          <year>2019</year>
          ). https://doi.org/10.1007/s10115-018-1220-z
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24.
          <string-name>
            <surname>Legg</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Smith</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Downing</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Visual analytics for collaborative human-machine confidence in human-centric active learning tasks</article-title>
          .
          <source>Hum. Cent. Comput. Inf. Sci. 9</source>
          (
          <issue>1</issue>
          ),
          <volume>5</volume>
          (
          <year>2019</year>
          ). https://doi.org/10.1186/s13673-019-0167-8
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          25.
          <string-name>
            <surname>Leite</surname>
            ,
            <given-names>R.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gschwandtner</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Miksch</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kriglstein</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pohl</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gstrein</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kuntner</surname>
          </string-name>
          , J.: Eva:
          <article-title>Visual analytics to identify fraudulent events</article-title>
          .
          <source>Trans. Vis. Comput. Graph</source>
          .
          <volume>24</volume>
          (
          <issue>1</issue>
          ),
          <fpage>330</fpage>
          -
          <lpage>339</lpage>
          (
          <year>2017</year>
          ). https://doi.org/10.1109/TVCG.
          <year>2017</year>
          .2744758
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          26.
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>Selecting training points for one-class support vector machines</article-title>
          .
          <source>Pattern Recognit. Lett</source>
          .
          <volume>32</volume>
          (
          <issue>11</issue>
          ),
          <fpage>1517</fpage>
          -
          <lpage>1522</lpage>
          (
          <year>2011</year>
          ). https://doi.org/10.1016/j.patrec.
          <year>2011</year>
          .
          <volume>04</volume>
          .013
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          27.
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gao</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gotz</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Du</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>He</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cao</surname>
          </string-name>
          , N.:
          <article-title>Rclens: Interactive rare category exploration and identification</article-title>
          .
          <source>Trans. Vis. Comput. Graph</source>
          .
          <volume>24</volume>
          (
          <issue>7</issue>
          ),
          <fpage>2223</fpage>
          -
          <lpage>2237</lpage>
          (
          <year>2017</year>
          ). https://doi.org/10.1109/TVCG.
          <year>2017</year>
          .2711030
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          28.
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Maljovec</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bremer</surname>
            ,
            <given-names>P.T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pascucci</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          :
          <article-title>Visualizing highdimensional data: Advances in the past decade</article-title>
          .
          <source>Trans. Vis. Comput. Graph</source>
          .
          <volume>23</volume>
          (
          <issue>3</issue>
          ),
          <fpage>1249</fpage>
          -
          <lpage>1268</lpage>
          (
          <year>2016</year>
          ). https://doi.org/10.1109/TVCG.
          <year>2016</year>
          .2640960
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          29.
          <string-name>
            <surname>Phillips</surname>
            ,
            <given-names>R.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chang</surname>
            ,
            <given-names>K.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Friedler</surname>
            ,
            <given-names>S.A.</given-names>
          </string-name>
          :
          <article-title>Interpretable active learning</article-title>
          .
          <source>arXiv preprint arXiv:1708.00049</source>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          30.
          <string-name>
            <surname>Ribeiro</surname>
            ,
            <given-names>M.T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Singh</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Guestrin</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          : “
          <string-name>
            <surname>Why Should I Trust You</surname>
          </string-name>
          <article-title>?”: Explaining the predictions of any classifier</article-title>
          .
          <source>In: SIGKDD</source>
          . pp.
          <fpage>1135</fpage>
          -
          <lpage>1144</lpage>
          (
          <year>2016</year>
          ). https://doi.org/10.1145/2939672.2939778
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          31.
          <string-name>
            <surname>Ringger</surname>
            ,
            <given-names>E.K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Carmen</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Haertel</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Seppi</surname>
            ,
            <given-names>K.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lonsdale</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>McClanahan</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Carroll</surname>
            ,
            <given-names>J.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ellison</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          :
          <article-title>Assessing the costs of machine-assisted corpus annotation through a user study</article-title>
          .
          <source>In: LREC</source>
          . pp.
          <fpage>3318</fpage>
          -
          <lpage>3324</lpage>
          (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          32.
          <string-name>
            <surname>Sacha</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sedlmair</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>J.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Peltonen</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Weiskopf</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          , North,
          <string-name>
            <given-names>S.C.</given-names>
            ,
            <surname>Keim</surname>
          </string-name>
          ,
          <string-name>
            <surname>D.A.</surname>
          </string-name>
          :
          <article-title>Visual interaction with dimensionality reduction: A structured literature analysis</article-title>
          .
          <source>Trans. Vis. Comput. Graph</source>
          .
          <volume>23</volume>
          (
          <issue>1</issue>
          ),
          <fpage>241</fpage>
          -
          <lpage>250</lpage>
          (
          <year>2016</year>
          ). https://doi.org/10.1109/TVCG.
          <year>2016</year>
          .2598495
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          33.
          <string-name>
            <surname>Seifert</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Granitzer</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>User-based active learning</article-title>
          .
          <source>In: ICDM Workshop</source>
          . pp.
          <fpage>418</fpage>
          -
          <lpage>425</lpage>
          . IEEE (
          <year>2010</year>
          ). https://doi.org/10.1109/ICDMW.
          <year>2010</year>
          .181
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          34.
          <string-name>
            <surname>Settles</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>From theories to queries: Active learning in practice</article-title>
          .
          <source>In: AISTATS Workshop</source>
          . pp.
          <fpage>1</fpage>
          -
          <lpage>18</lpage>
          (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          35.
          <string-name>
            <surname>Settles</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Active learning</article-title>
          .
          <source>Synthesis Lectures on Artificial Intelligence and Machine</source>
          Learning pp.
          <fpage>1</fpage>
          -
          <lpage>114</lpage>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          36.
          <string-name>
            <surname>Sperrle</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bernard</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sedlmair</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Keim</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>El-Assady</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Speculative execution for guided visual analytics</article-title>
          .
          <source>In: VIS Workshop</source>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          37.
          <string-name>
            <surname>Stokes</surname>
            ,
            <given-names>J.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Platt</surname>
            ,
            <given-names>J.C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kravis</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shilman</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Aladin: Active learning of anomalies to detect intrusions</article-title>
          .
          <source>Tech. rep., Microsoft Research</source>
          (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          38.
          <string-name>
            <surname>Sun</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Qu</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Di</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gao</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>Heuristic sample reduction method for support vector data description</article-title>
          .
          <source>Turkish Journal of Electrical Engineering &amp; Computer Sciences</source>
          <volume>24</volume>
          (
          <issue>1</issue>
          ),
          <fpage>298</fpage>
          -
          <lpage>312</lpage>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          39.
          <string-name>
            <surname>Tax</surname>
            ,
            <given-names>D.M.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Duin</surname>
            ,
            <given-names>R.P.W.</given-names>
          </string-name>
          :
          <article-title>Support vector data description</article-title>
          .
          <source>Mach. Learn</source>
          .
          <volume>54</volume>
          (
          <issue>1</issue>
          ),
          <fpage>45</fpage>
          -
          <lpage>66</lpage>
          (
          <year>2004</year>
          ). https://doi.org/10.1023/B:MACH.
          <volume>0000008084</volume>
          .60811.49
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          40.
          <string-name>
            <surname>Teso</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kersting</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>Explanatory interactive machine learning</article-title>
          .
          <source>In: AAAI</source>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref41">
        <mixed-citation>
          41.
          <string-name>
            <surname>Trittenbach</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Englhardt</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , Bo¨hm,
          <string-name>
            <surname>K.</surname>
          </string-name>
          :
          <article-title>An overview and a benchmark of active learning for outlier detection with One-Class classifiers</article-title>
          . arXiv:
          <year>1808</year>
          .
          <volume>04759</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref42">
        <mixed-citation>
          42.
          <string-name>
            <surname>Vollmer</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Englhardt</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Trittenbach</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bielski</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Karrari</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , Bo¨hm, K.:
          <article-title>Energy time-series features for emerging applications on the basis of human-readable machine descriptions</article-title>
          .
          <source>In: Proceedings of the Tenth ACM International Conference on Future Energy Systems</source>
          . pp.
          <fpage>474</fpage>
          -
          <lpage>481</lpage>
          . ACM (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref43">
        <mixed-citation>
          43.
          <string-name>
            <surname>Wilkinson</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Visualizing big data outliers through distributed aggregation</article-title>
          .
          <source>Trans. Vis. Comput. Graph</source>
          .
          <volume>24</volume>
          (
          <issue>1</issue>
          ),
          <fpage>256</fpage>
          -
          <lpage>266</lpage>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref44">
        <mixed-citation>
          44.
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Loog</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>A benchmark and comparison of active learning for logistic regression</article-title>
          .
          <source>Pattern Recognit</source>
          .
          <volume>83</volume>
          ,
          <fpage>401</fpage>
          -
          <lpage>415</lpage>
          (
          <year>2018</year>
          ). https://doi.org/10.1109/TVCG.
          <year>2017</year>
          .2744685
        </mixed-citation>
      </ref>
      <ref id="ref45">
        <mixed-citation>
          45.
          <string-name>
            <surname>Yin</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fan</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          :
          <article-title>Active learning based support vector data description method for robust novelty detection</article-title>
          .
          <source>Knowl. Based. Syst</source>
          .
          <volume>153</volume>
          ,
          <fpage>40</fpage>
          -
          <lpage>52</lpage>
          (
          <year>2018</year>
          ). https://doi.org/10.1016/j.knosys.
          <year>2018</year>
          .
          <volume>04</volume>
          .020
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>