<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>PhD Workshop, August</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Unsupervised Ensembles for Outlier Detection</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>Guilherme O. Campos Supervised by Prof. Dr. Wagner Meira Jr. Department of Computer Science Federal University of Minas Gerais Belo Horizonte</institution>
          ,
          <country country="BR">Brazil</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2018</year>
      </pub-date>
      <volume>27</volume>
      <issue>2018</issue>
      <abstract>
        <p>Ensemble techniques have been applied to the unsupervised outlier detection problem in some scenarios. Challenges are the generation of diverse ensemble members and the combination of individual results into an ensemble. For the latter challenge, some methods tried to design smaller ensembles out of a wealth of possible ensemble members, to improve the diversity and accuracy of the ensemble (relating to the ensemble selection problem in classi cation). In this paper, We propose a boosting strategy to solve the ensemble selection problem, called BoostSelect. We evaluate BoostSelect over a large benchmark of datasets for outlier detection, showing improvements over baseline approaches.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. INTRODUCTION</title>
      <p>The identi cation of outliers (i.e., data objects that do not
t well to the general data distribution) is very important in
many practical applications. Application examples are the
detection of credit card fraud in nancial transactions data,
the identi cation of measurement errors in scienti c data,
or the analysis of sports statistics data.</p>
      <p>
        Recent research on the unsupervised problem of outlier
detection advanced the area by applying ensemble techniques
[
        <xref ref-type="bibr" rid="ref31">31</xref>
        ]. Ensemble methods, i.e., combining the ndings or
results of individual learners to an integrated, typically more
reliable and better result, are well established in the
supervised context of classi cation or regression [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ]. In
unsupervised learning, the theoretical underpinnings are less clear
but can be drawn in analogy to the supervised context as it
has been done for clustering ensembles [
        <xref ref-type="bibr" rid="ref30">30</xref>
        ].
      </p>
      <p>
        The focus of my PhD thesis is on ensemble selection,
which has been well studied in supervised scenarios [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] (also
called selective ensembles [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ], or ensemble pruning [
        <xref ref-type="bibr" rid="ref15 ref27 ref30">15, 27,
30</xref>
        ]). Ensemble selection is also related to boosting [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ],
which is often used to change training conditions for
additionally sought, yet to be trained ensemble members or to
select the most suitable additional ensemble members from
a pool of solutions.
      </p>
      <p>Given many outlier detection results, how do we select
which members we are going to include into the ensemble?
How do we combine these selected members into a single and
more robust result? These challenges mixed to an
unsupervised environment are still pertinent and more research is
need to increase our knowledge over this topic in question.</p>
      <p>In this paper we propose an initial approach towards a
more robust ensemble method by transferring the supervised
boosting technique to the unsupervised scenario of outlier
detection ensembles. The proposed outlier ensemble
selection technique is called BoostSelect.
2.</p>
    </sec>
    <sec id="sec-2">
      <title>RELATED WORK</title>
      <p>
        The ensemble approach to learning has been studied in
outlier detection several times. In analogy to supervised
learning, an ensemble can be expected to improve over its
components if these components deliver results with a
reasonable accuracy while being diverse [
        <xref ref-type="bibr" rid="ref31">31</xref>
        ]. The two main
challenges for creating good ensembles are, therefore, (i) the
generation of diverse (potential) ensemble members, and (ii)
the combination (or selection) of members to an ensemble.
      </p>
      <p>
        Some strategies to achieve diversity among ensemble
members have been explored, such as feature bagging (i.e.,
combining outlier scores learned on di erent subsets of attributes)
[
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], di erent parameter choices for some base method [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ],
the combination of actually di erent base methods [
        <xref ref-type="bibr" rid="ref10 ref16 ref23">16, 10,
23</xref>
        ], the introduction of a random component in a given
learner [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], the use of di erent subsamples of the data
objects (parallel [
        <xref ref-type="bibr" rid="ref33">33</xref>
        ] or sequential [
        <xref ref-type="bibr" rid="ref19 ref21">21, 19</xref>
        ]), adding some
random noise component on the data (\perturbation") [
        <xref ref-type="bibr" rid="ref32">32</xref>
        ], or
using approximate neighborhoods for density estimates [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
Likewise, di erent combination procedures have been
proposed based on outlier scores or on outlier ranks [
        <xref ref-type="bibr" rid="ref10 ref13 ref31 ref5">13, 5, 10,
31</xref>
        ].
      </p>
      <p>
        Some methods have also been proposed to select the more
diverse or the more accurate ensemble members [
        <xref ref-type="bibr" rid="ref18 ref23">23, 18</xref>
        ].
These unsupervised methods construct a target result vector
from un ltered results and then sequentially select
individual results that somehow t to the target vector while
being di erent from already selected solutions. The \Greedy
ensemble" approach [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ] fails in generating a good target
vector by selecting a percentage of top instances for each
ensemble candidate to compose the vector. This procedure
normally selects many inliers to t the target vector due
to the imbalance between few outliers and many inliers. To
solve this problem, \SelectV" approach [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] generate the
target vector by taking the average over all outlier detection
results (possible ensemble members). The main problem with
\SelectV" is that the algorithm does not consider diversity
as an important factor in selecting ensemble members.
      </p>
    </sec>
    <sec id="sec-3">
      <title>BOOSTING FOR ENSEMBLE SELECTION</title>
      <p>
        Starting from the ideas discussed for the \Greedy
ensemble" [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ] and for \SelectV" [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ], we propose here an
improved outlier ensemble selection method that is amenable
to the application of boosting techniques. Boosting is well
studied in supervised contexts [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ]. We design and apply
an equivalent technique in the unsupervised setting, to
select good components for an ensemble of outlier detectors,
resulting in our method BoostSelect.
3.1
      </p>
    </sec>
    <sec id="sec-4">
      <title>Construction of the Target Vector</title>
      <p>
        As a prerequisite for the combination of di erent outlier
score lists (i.e., individual results, potential ensemble
members), we normalize the scores following established
procedures [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. The target vector is constructed by combining
the scores of all available results. Di erent combination
methods could be used here, without further assumptions
taking the average score is the most natural approach [
        <xref ref-type="bibr" rid="ref31">31</xref>
        ],
i.e., the target vector lists the average scores of all
individual results for each data object. From this target vector, we
preliminarily assume the top bn tc objects (ranked by their
combined score) to be outliers, where n is the dataset size
and 0 &lt; t 1 is a parameter capturing the expected
percentage of outliers in the dataset (i.e., there are K = bn tc
outliers assumed to be present). The target vector thus
becomes a binary vector, listing 1 for an (alleged) outlier and
0 for an (alleged) inlier and serves as pseudo ground truth
for the boosting approach to ensemble selection.
3.2
      </p>
    </sec>
    <sec id="sec-5">
      <title>Weights and Ensemble Diversity</title>
      <p>
        Weighted Pearson correlation has been proposed as a
similarity measure for outlier rankings [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ]. We follow the
procedure of Schubert et al. [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ], setting weights for Pearson
correlation to outliers and inliers. Di erent from previous
approaches, though, these values are only the initial weights.
The weights will be updated by the boosting procedure.
      </p>
      <p>The potential ensemble members are sorted according to
their weighted Pearson correlation to the target vector. The
candidate that is most similar to the target vector is chosen
as the rst ensemble member.</p>
      <p>Remaining potential ensemble members are iteratively
resorted in ascending order according to their similarity to the
current prediction of the ensemble, resulting in a preference
for the most di erent (i.e., most complementary) additional
ensemble members. Potential members are included if their
inclusion would increase the similarity of the ensemble
prediction to the target vector, otherwise they are discarded.
If the correlation improves, the ensemble is updated and
the remaining lists are re-sorted by their weighted Pearson
correlation to the updated prediction.
3.3</p>
    </sec>
    <sec id="sec-6">
      <title>Boosting Procedure</title>
      <p>The boosting is performed upon the inclusion of a new
member into the ensemble. The idea is to reduce the weights
for those outliers that have already been identi ed by any
ensemble member. The weights are reduced by some
specied parameter 0 &lt; d &lt; 1 (drop rate).</p>
      <p>The boosting e ect is that the selection will prefer to
include such additional ensemble members that detect those
outliers that have not yet been detected by any ensemble</p>
      <p>Algorithm 1 BoostSelect
Input: P := set of normalized outlier score lists,
d := drop rate (percentage), t := threshold (percentage),
combination := combination technique
Output: E := ensemble members
1: W := [n]; E := ;
2: target := combination(P ) . Generating the target
vector
3: target := convertBinary(target; t) . Top K = bn tc
scores 1, others 0
4: W := hout = 21K ; in = 2(n1 K)
i
. K = number of
outliers, n = size
5: Sort P by weighted Pearson Correlation (wP C) to
target . Descending order
6: f := getF irst(P ) . Remove f from P
7: E := E [ f
8: while P 6= ; do
9: curr := combination(E) . Current prediction
10: sort P by wP C to curr . Ascending order
11: f := getF irst(P ) . Remove f from P
12: if wP C(combination(E [ f ); target) &gt;
wP C(curr; target) then
13: E := E [ f . Include into ensemble
14: Boosting(W; target; f; t; d) . Adapt the weights
15: end if
16: end while
Algorithm 2 Boosting
Input: W := weight vector, target := target vector, f :=
new ensemble member, t := threshold (percentage), d :=
drop rate (percentage)
Output: W := Updated weights
1: outliers := convertBinary(f; t)
2: for i 2 1 : size(target) do
3: if target(i) == 1 &amp; outliers(i) == 1 then
4: W (i) := W (i) d
5: end if
6: end for
member, while very easy outliers that have been detected by
many ensemble members already will get assigned smaller
and smaller weights.</p>
      <p>Algorithm 1 lists the steps of the overall framework
BoostSelect in pseudo code. The boosting procedure is detailed
in Algorithm 2.
4.
4.1</p>
    </sec>
    <sec id="sec-7">
      <title>EXPERIMENTS</title>
    </sec>
    <sec id="sec-8">
      <title>Datasets</title>
      <p>
        For evaluation, we use a benchmark data repository for
outlier detection [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. The repository is based on 23 basic
datasets, processed in di erent ways mainly to provide
variants with di erent percentage of outliers and with di erent
handling of dataset characteristics such as duplicates,
attribute normalization, and categorical values. As suggested
for analysis [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], we focus on the normalized datasets without
duplicates, which leaves us with 422 dataset variants.
4.2
      </p>
    </sec>
    <sec id="sec-9">
      <title>Ensemble Members</title>
      <p>
        As basic outlier detection results (i.e., potential
ensemble members) we use the results provided along with the
datasets [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], testing 12 neighborhood-based outlier
detection algorithms changing the neighborhood size k from 1 to
100. The algorithms are: KNN [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], KNNW [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], LOF [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ],
Simpli edLOF [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ], LoOP [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], LDOF [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ], ODIN [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ],
FastABOD [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], KDEOS [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ], LDF [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], INFLO [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], and COF [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ].
For LDOF and KDEOS, k must be larger than 1, for
FastABOD, k must be larger than 2, resulting in 1196 results per
dataset (less on some small datasets where k cannot reach
100). These results compose the set of potential ensemble
members.
      </p>
      <p>
        The outlier scores of these results are processed (following
Kriegel et al. [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]) by applying an inverse logarithmic scaling
on FastABOD results and an inverse linear scaling on ODIN
results, since FastABOD and ODIN give inverse score results
(i.e., the lower the scores, the higher is the chance of an
observation to be an outlier). Then a simple linear scaling
from 0 to 1 is applied to transform all scores into the same
range.
      </p>
    </sec>
    <sec id="sec-10">
      <title>4.3 Competitors and Settings</title>
      <p>
        We compare BoostSelect against the Greedy [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ] and
SelectV [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] ensembles. We also generate a \Nave" ensemble
and Random ensembles as baselines. The \Nave" ensemble
is a combination of all individual outlier results (i.e., a full
ensemble without selection procedure).
      </p>
      <p>For each instance of an ensemble selection strategy (Greedy,
SelectV, and BoostSelect, respectively, on each dataset), we
generate 1000 \Random" ensembles consisting of the same
number of members as the corresponding selective ensemble,
where the ensemble members are randomly selected.</p>
      <p>
        We used the Greedy ensemble rate parameter as 0:01, as
suggested by the authors of the Greedy ensemble [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ]. We
test a range of parameters for BoostSelect: d = [0:25; 0:5; 0:75]
and t = [0:05; 0:1; 0:15].
      </p>
      <p>As combination technique for ensembles we use the
average score.</p>
    </sec>
    <sec id="sec-11">
      <title>4.4 Results</title>
      <p>Figure 1 shows pairwise comparisons between all
ensembles over all datasets, considering the ROC AUC evaluation
measure (area under the curve of the receiver operating
characteristic). We compare the ensemble selection techniques
\Nave", \Greedy", \SelectV", and \BS" (BoostSelect). We
include random ensembles for each ensemble selection
strategy and for each parametrization of BoostSelect: RG
(Random Greedy), RS (Random SelectV), RBS (Random
BoostSelect). The numbers represent on how many datasets the
ensemble listed in the row has performed better than the
ensemble listed in the column. Numbers representing the
majority (more than 50%) of the datasets are white, smaller
numbers black. The larger the number, the darker is its
background. For the random ensemble, we take the average
performance over the 1000 instances.</p>
      <p>The best overall method is BoostSelect with d = 0:75 and
t = 0:05, which has only more losses than wins when
competing against BoostSelect with d = 0:25 and t = 0:1. The
Greedy ensemble does not perform well in general, having
more losses than wins against every other competitor.
SelectV is better than all random variants and Greedy, but
worse than Nave and worse than all BoostSelect results.
The Nave ensemble behaves very consistently, as it beats by
a large margin all random ensemble approaches, but still has
more losses when compared to BoostSelect. Even though
neither the threshold t nor the drop rate d has a strong
impact on wins, setting a relatively large drop rate and a
relatively small threshold overall seems to be a good choice
of parameters for BoostSelect, although the optimal
parameter choice di ers from dataset to dataset.</p>
      <p>Looking at the top left quadrant of the heat map
(Figure 1), where the random ensembles compete against
themselves, we also see a broad dominance by the random
ensembles based on BoostSelect. This suggests that the number of
ensemble members selected by BoostSelect is a better choice
than those selected by the other strategies.
300
250
200
150
100
50
0</p>
    </sec>
    <sec id="sec-12">
      <title>5. CONCLUSION AND FUTURE DIRECTIONS</title>
      <p>We proposed a new ensemble selection strategy for
unsupervised outlier detection ensembles, using the unsupervised
equivalent to a boosting strategy for ensemble selection.
Experiments show the favorable behavior of the new ensemble
selection strategy compared to existing methods (Greedy
and SelectV) on a large set of benchmark datasets. Main
di erences between our method BoostSelect, the Greedy
ensemble, and SelectV can be attributed to a di erent way of
focusing on diversity and accuracy of ensemble members.
Greedy goes all out for diversity and mostly disregards
accuracy, while SelectV ignores diversity and maximizes
accuracy of the ensemble members. Our new method
BoostSelect considers both, diversity and accuracy, in a balanced
manner and performs competitively on average over a large
selection of benchmark datasets with strong improvements
on many of the benchmark datasets.</p>
      <p>The behavior of BoostSelect is robust to the parameters
on many datasets but depends strongly on the choice of
parameters on some datasets. As future work, we are specially
interested on this behavior and potential relation to
properties of the datasets. We are also interested to improve
the target vector generation step and to include a complete
study over the combination step of outlier detection
ensemble.</p>
      <p>Acknowledgments This work was partially supported by
CAPES - Brazil, Fapemig, CNPq, and by projects InWeb,
MASWeb, EUBra-BIGSEA, INCT-Cyber, and Atmosphere.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>F.</given-names>
            <surname>Angiulli</surname>
          </string-name>
          and
          <string-name>
            <given-names>C.</given-names>
            <surname>Pizzuti</surname>
          </string-name>
          .
          <article-title>Fast outlier detection in high dimensional spaces</article-title>
          .
          <source>In Proc. PKDD</source>
          , pages
          <volume>15</volume>
          {
          <fpage>26</fpage>
          ,
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>M. M. Breunig</surname>
            ,
            <given-names>H.-P.</given-names>
          </string-name>
          <string-name>
            <surname>Kriegel</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Ng</surname>
            , and
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Sander</surname>
          </string-name>
          . LOF:
          <article-title>Identifying density-based local outliers</article-title>
          .
          <source>In Proc. SIGMOD</source>
          , pages
          <volume>93</volume>
          {
          <fpage>104</fpage>
          ,
          <year>2000</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>G. O.</given-names>
            <surname>Campos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Zimek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Sander</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. J. G. B.</given-names>
            <surname>Campello</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Micenkov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Schubert</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Assent</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M. E.</given-names>
            <surname>Houle</surname>
          </string-name>
          .
          <article-title>On the evaluation of unsupervised outlier detection: Measures, datasets, and an empirical study</article-title>
          .
          <source>Data Min. Knowl. Disc.</source>
          ,
          <volume>30</volume>
          :
          <fpage>891</fpage>
          {
          <fpage>927</fpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>R.</given-names>
            <surname>Caruana</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Niculescu-Mizil</surname>
          </string-name>
          ,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <article-title>Crew, and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Ksikes</surname>
          </string-name>
          .
          <article-title>Ensemble selection from libraries of models</article-title>
          .
          <source>In Proc. ICML</source>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J.</given-names>
            <surname>Gao</surname>
          </string-name>
          and
          <string-name>
            <given-names>P.-N.</given-names>
            <surname>Tan</surname>
          </string-name>
          .
          <article-title>Converting output scores from outlier detection algorithms into probability estimates</article-title>
          .
          <source>In Proc. ICDM</source>
          , pages
          <volume>212</volume>
          {
          <fpage>221</fpage>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>V.</given-names>
            <surname>Hautamki</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Krkkinen</surname>
          </string-name>
          , and
          <string-name>
            <given-names>P.</given-names>
            <surname>Frnti</surname>
          </string-name>
          .
          <article-title>Outlier detection using k-nearest neighbor graph</article-title>
          .
          <source>In Proc. ICPR</source>
          , pages
          <volume>430</volume>
          {
          <fpage>433</fpage>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>W.</given-names>
            <surname>Jin</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. K. H. Tung</surname>
            , J. Han, and
            <given-names>W.</given-names>
          </string-name>
          <string-name>
            <surname>Wang</surname>
          </string-name>
          .
          <article-title>Ranking outliers using symmetric neighborhood relationship</article-title>
          .
          <source>In Proc. PAKDD</source>
          , pages
          <volume>577</volume>
          {
          <fpage>593</fpage>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>E.</given-names>
            <surname>Kirner</surname>
          </string-name>
          , E. Schubert,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Zimek</surname>
          </string-name>
          .
          <article-title>Good and bad neighborhood approximations for outlier detection ensembles</article-title>
          .
          <source>In Proc. SISAP</source>
          , pages
          <volume>173</volume>
          {
          <fpage>187</fpage>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>H.-P.</given-names>
            <surname>Kriegel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Krger</surname>
          </string-name>
          , E. Schubert,
          <article-title>and</article-title>
          <string-name>
            <surname>A. Zimek.</surname>
          </string-name>
          <article-title>LoOP: local outlier probabilities</article-title>
          .
          <source>In Proc. CIKM</source>
          , pages
          <volume>1649</volume>
          {
          <fpage>1652</fpage>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>H.-P.</given-names>
            <surname>Kriegel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Krger</surname>
          </string-name>
          , E. Schubert,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Zimek</surname>
          </string-name>
          .
          <article-title>Interpreting and unifying outlier scores</article-title>
          .
          <source>In Proc. SDM</source>
          , pages
          <volume>13</volume>
          {
          <fpage>24</fpage>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>H.-P.</given-names>
            <surname>Kriegel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Schubert</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Zimek</surname>
          </string-name>
          .
          <article-title>Angle-based outlier detection in high-dimensional data</article-title>
          .
          <source>In Proc. KDD</source>
          , pages
          <volume>444</volume>
          {
          <fpage>452</fpage>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>L. J.</given-names>
            <surname>Latecki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Lazarevic</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D.</given-names>
            <surname>Pokrajac</surname>
          </string-name>
          .
          <article-title>Outlier detection with kernel density functions</article-title>
          .
          <source>In Proc. MLDM</source>
          , pages
          <volume>61</volume>
          {
          <fpage>75</fpage>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>A.</given-names>
            <surname>Lazarevic</surname>
          </string-name>
          and
          <string-name>
            <given-names>V.</given-names>
            <surname>Kumar</surname>
          </string-name>
          .
          <article-title>Feature bagging for outlier detection</article-title>
          .
          <source>In Proc. KDD</source>
          , pages
          <volume>157</volume>
          {
          <fpage>166</fpage>
          ,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>F. T.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. M.</given-names>
            <surname>Ting</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Z.-H.</given-names>
            <surname>Zhou</surname>
          </string-name>
          .
          <article-title>Isolation-based anomaly detection</article-title>
          .
          <source>ACM TKDD</source>
          ,
          <volume>6</volume>
          (
          <issue>1</issue>
          ):3:1{
          <fpage>39</fpage>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>D. D.</given-names>
            <surname>Margineantu</surname>
          </string-name>
          and
          <string-name>
            <given-names>T. G.</given-names>
            <surname>Dietterich</surname>
          </string-name>
          .
          <article-title>Pruning adaptive boosting</article-title>
          .
          <source>In Proc. ICML</source>
          , pages
          <volume>211</volume>
          {
          <fpage>218</fpage>
          ,
          <year>1997</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>H. V.</given-names>
            <surname>Nguyen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. H.</given-names>
            <surname>Ang</surname>
          </string-name>
          , and
          <string-name>
            <given-names>V.</given-names>
            <surname>Gopalkrishnan</surname>
          </string-name>
          .
          <article-title>Mining outliers with ensemble of heterogeneous detectors on random subspaces</article-title>
          .
          <source>In Proc. DASFAA</source>
          , pages
          <volume>368</volume>
          {
          <fpage>383</fpage>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>S.</given-names>
            <surname>Ramaswamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Rastogi</surname>
          </string-name>
          , and
          <string-name>
            <given-names>K.</given-names>
            <surname>Shim</surname>
          </string-name>
          .
          <article-title>E cient algorithms for mining outliers from large data sets</article-title>
          .
          <source>In Proc. SIGMOD</source>
          , pages
          <volume>427</volume>
          {
          <fpage>438</fpage>
          ,
          <year>2000</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>S.</given-names>
            <surname>Rayana</surname>
          </string-name>
          and
          <string-name>
            <given-names>L.</given-names>
            <surname>Akoglu</surname>
          </string-name>
          .
          <article-title>Less is more: Building selective anomaly ensembles</article-title>
          .
          <source>ACM TKDD</source>
          ,
          <volume>10</volume>
          (
          <issue>4</issue>
          ):
          <volume>42</volume>
          :1{
          <fpage>42</fpage>
          :
          <fpage>33</fpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>S.</given-names>
            <surname>Rayana</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Zhong</surname>
          </string-name>
          , and
          <string-name>
            <given-names>L.</given-names>
            <surname>Akoglu</surname>
          </string-name>
          .
          <article-title>Sequential ensemble learning for outlier detection: A bias-variance perspective</article-title>
          .
          <source>In Proc. ICDM</source>
          , pages
          <volume>1167</volume>
          {
          <fpage>1172</fpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>L.</given-names>
            <surname>Rokach</surname>
          </string-name>
          .
          <article-title>Ensemble-based classi ers</article-title>
          .
          <source>Artif. Intell. Rev.</source>
          ,
          <volume>33</volume>
          :1{
          <fpage>39</fpage>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>M.</given-names>
            <surname>Salehi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. C.</given-names>
            <surname>Bezdek</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C.</given-names>
            <surname>Leckie</surname>
          </string-name>
          .
          <article-title>Smart sampling: A novel unsupervised boosting approach for outlier detection</article-title>
          .
          <source>In AI 2016: Advances in Arti cial Intelligence - 29th Australasian Joint Conference</source>
          , Hobart,
          <string-name>
            <surname>TAS</surname>
          </string-name>
          , Australia, December 5-
          <issue>8</issue>
          ,
          <year>2016</year>
          , Proceedings, pages
          <volume>469</volume>
          {
          <fpage>481</fpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>R. E.</given-names>
            <surname>Schapire</surname>
          </string-name>
          and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Freund</surname>
          </string-name>
          . Boosting.
          <article-title>Foundations and Algorithms</article-title>
          . MIT Press, Cambridge, MA,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>E.</given-names>
            <surname>Schubert</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Wojdanowski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Zimek</surname>
          </string-name>
          , and
          <string-name>
            <given-names>H.-P.</given-names>
            <surname>Kriegel</surname>
          </string-name>
          .
          <article-title>On evaluation of outlier rankings and outlier scores</article-title>
          .
          <source>In Proc. SDM</source>
          , pages
          <volume>1047</volume>
          {
          <fpage>1058</fpage>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>E.</given-names>
            <surname>Schubert</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Zimek</surname>
          </string-name>
          , and
          <string-name>
            <given-names>H.-P.</given-names>
            <surname>Kriegel</surname>
          </string-name>
          .
          <article-title>Generalized outlier detection with exible kernel density estimates</article-title>
          .
          <source>In Proc. SDM</source>
          , pages
          <volume>542</volume>
          {
          <fpage>550</fpage>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>E.</given-names>
            <surname>Schubert</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Zimek</surname>
          </string-name>
          , and
          <string-name>
            <given-names>H.-P.</given-names>
            <surname>Kriegel</surname>
          </string-name>
          .
          <article-title>Local outlier detection reconsidered: a generalized view on locality with applications to spatial, video, and network outlier detection</article-title>
          .
          <source>Data Min. Knowl. Disc.</source>
          ,
          <volume>28</volume>
          (
          <issue>1</issue>
          ):
          <volume>190</volume>
          {
          <fpage>237</fpage>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>J.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. W.-C.</given-names>
            <surname>Fu</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D. W.</given-names>
            <surname>Cheung</surname>
          </string-name>
          .
          <article-title>Enhancing e ectiveness of outlier detections for low density patterns</article-title>
          .
          <source>In Proc. PAKDD</source>
          , pages
          <volume>535</volume>
          {
          <fpage>548</fpage>
          ,
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>G.</given-names>
            <surname>Tsoumakas</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Partalas</surname>
          </string-name>
          ,
          <string-name>
            <surname>and I. Vlahavas.</surname>
          </string-name>
          <article-title>An ensemble pruning primer</article-title>
          . In O. Okun and G. Valentini, editors,
          <source>Applications of Supervised and Unsupervised Ensemble Methods</source>
          , pages
          <fpage>1</fpage>
          <lpage>{</lpage>
          13. Springer,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>K.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hutter</surname>
          </string-name>
          , and
          <string-name>
            <given-names>H.</given-names>
            <surname>Jin</surname>
          </string-name>
          .
          <article-title>A new local distance-based outlier detection approach for scattered real-world data</article-title>
          .
          <source>In Proc. PAKDD</source>
          , pages
          <volume>813</volume>
          {
          <fpage>822</fpage>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wu</surname>
          </string-name>
          , and
          <string-name>
            <given-names>W.</given-names>
            <surname>Tang</surname>
          </string-name>
          .
          <article-title>Ensembling neural networks: Many could be better than all</article-title>
          .
          <source>Artif</source>
          . Intell.,
          <volume>137</volume>
          (
          <issue>1-2</issue>
          ):
          <volume>239</volume>
          {
          <fpage>263</fpage>
          ,
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>Z.-H.</given-names>
            <surname>Zhou</surname>
          </string-name>
          . Ensemble Methods.
          <article-title>Foundations and Algorithms</article-title>
          . CRC Press,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>A.</given-names>
            <surname>Zimek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. J. G. B.</given-names>
            <surname>Campello</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.</given-names>
            <surname>Sander</surname>
          </string-name>
          .
          <article-title>Ensembles for unsupervised outlier detection: Challenges and research questions</article-title>
          .
          <source>SIGKDD Explor</source>
          .,
          <volume>15</volume>
          (
          <issue>1</issue>
          ):
          <volume>11</volume>
          {
          <fpage>22</fpage>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32]
          <string-name>
            <given-names>A.</given-names>
            <surname>Zimek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. J. G. B.</given-names>
            <surname>Campello</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.</given-names>
            <surname>Sander</surname>
          </string-name>
          .
          <article-title>Data perturbation for outlier detection ensembles</article-title>
          .
          <source>In Proc. SSDBM</source>
          , pages
          <volume>13</volume>
          :1{
          <fpage>12</fpage>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [33]
          <string-name>
            <given-names>A.</given-names>
            <surname>Zimek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Gaudet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. J. G. B.</given-names>
            <surname>Campello</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.</given-names>
            <surname>Sander</surname>
          </string-name>
          .
          <article-title>Subsampling for e cient and e ective unsupervised outlier detection ensembles</article-title>
          .
          <source>In Proc. KDD</source>
          , pages
          <volume>428</volume>
          {
          <fpage>436</fpage>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>