<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Anomaly Detection Using Unsupervised Profiling Method in Time Series Data</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Zakia Ferdousi</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Akira Maeda</string-name>
          <email>amaeda@is.ritsumei.ac.jp</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Media Technology, College of Information Science and Engineering, Ritsumeikan University</institution>
          ,
          <addr-line>1-1-1, Noji-Higashi, Kusatsu, Shiga, 525-8577</addr-line>
          ,
          <country country="JP">Japan</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Graduate School of Science and Engineering, Ritsumeikan University</institution>
          ,
          <addr-line>1-1-1, Noji-Higashi, Kusatsu, Shiga, 525-8577</addr-line>
          ,
          <country country="JP">Japan</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The anomaly detection problem has important applications in the field of fraud detection, network robustness analysis and intrusion detection. This paper is concerned with the problem of detecting anomalies in time series data using Peer Group Analysis (PGA), which is an unsupervised technique. The objective of PGA is to characterize the expected pattern of behavior around the target sequence in terms of the behavior of similar objects and then to detect any differences in evolution between the expected pattern and the target. The experimental results demonstrate that the method is able to flag anomalous records effectively.</p>
      </abstract>
      <kwd-group>
        <kwd>Anomaly Detection</kwd>
        <kwd>Data Mining</kwd>
        <kwd>Peer Group Analysis</kwd>
        <kwd>Unsupervised Profiling</kwd>
        <kwd>Time Series Data</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1 Introduction</title>
      <p>
        With the expanded Internet and the increase of online financial transactions, financial
services companies have become more vulnerable to fraud. Outlier detection is a
fundamental issue in data mining, specifically in fraud detection. Outliers have been
informally defined as observations in a data set which appear to be inconsistent with the
remainder of that set of data [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ], or which deviate so much from other observations
so as to arouse suspicions that they were generated by a different mechanism [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. The
identification of outliers can lead to the discovery of useful knowledge and has a
number of practical applications in areas such as credit card fraud detection, athlete
performance analysis, voting irregularity analysis, severe weather prediction etc. [
        <xref ref-type="bibr" rid="ref4 ref5 ref6">4, 5,
6</xref>
        ]. Peer Group Analysis (PGA) is an unsupervised method for monitoring behavior
over time in data mining [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Unsupervised methods do not need the prior knowledge
of fraudulent and non-fraudulent transactions in historical database, but instead detect
changes in behavior or unusual transactions.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2 Stock Market Analysis</title>
      <sec id="sec-2-1">
        <title>2.1 Stock Fraud and the manipulators</title>
        <p>Stock fraud usually takes place when brokers try to manipulate their customers into
trading stocks without regard for the customers' own real interests. Corporate insiders,
brokers, underwriters, large shareholders and market makers are likely to be
manipulators.
2.2</p>
      </sec>
      <sec id="sec-2-2">
        <title>Why Stock Fraud Detection is Necessary?</title>
        <p>Several fraud detection methods are available for the fields like credit card,
telecommunications, network intrusion detections etc. But stock market fraud detection area
is still lagging. Since stock market enhances the economic development of a country
greatly, this field has a vital need for efficient security system. Also the amount of
money involved in stock market is huge. For example, in Australia, 63 per cent of
people's retirement savings is invested in securities. Investment in stock market is
high in almost all the countries. If we do not protect against the ability of people to
manipulate those securities, then implicitly, we are open to attack, or we are allowing
open to attack a country's wealth indeed. It is a very real threat, a threat that very few
people really, are acknowledging. Stock fraud may not be very frequent but when it
arises the amount of loss is abundant. Outlier detection in stock market transactions
will not only prevent the fraud but also alert the stock markets and broking houses to
unusual movements in the markets.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3 Our Contribution</title>
      <p>First we analyzed how fraudulent cases occur in stock market through the thorough
technical reviews and from the practical experiences of stock markets. The following
two cases are the most important criteria, which we aim to mine first to detect the
stock fraud:
</p>
      <p>To identify broker accounts whose sell quantity rise up or fall down
suddenly.
 To identify broker accounts whose trade volume rise up or fall down
suddenly.</p>
      <p>
        We simulate the PGA tool in various situations and illustrate its use on a set of
stock market transaction data. PGA was initially proposed for credit card fraud
detection by Bolton &amp; Hand in 2001[
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], where it considered only the spending amount of
each account. But using one attribute is not enough to flag an account as a fraud. An
effective and practical fraud detection method needs to incorporate more information.
We tried to overcome the problem by including more attributes within the outlier
detection process by PGA. We applied PGA over two attributes and then we performed
a comparative analysis between those two observations. We flagged the accounts as
suspicious based on the knowledge discovered from the comparative analysis. Thus
the results of outliers mining become more realistic and effective than the traditional
PGA. We also demonstrated t-statistics to find the deviations more effectively.
      </p>
    </sec>
    <sec id="sec-4">
      <title>4 Related Work</title>
      <p>
        Outlier detection in time series database has recently received considerable attention
in the field of data mining. Qu, et al. uses probabilities of events to define the profile
[
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], Lane and Brodley [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], Forrest et al. [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] and Kosoresow and Hofmeyr [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] use
similarity of sequences that can be interpreted in a probabilistic framework.
      </p>
      <p>
        The neural network and Bayesian network comparison study [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] uses the STAGE
algorithm for Bayesian networks and back propagation algorithm for neural networks
in credit transactional fraud detection. Comparative results show that Bayesian
networks were more accurate and much faster to train, but Bayesian networks are slower
when applied to new instances. The Securities Observation, News Analysis, and
Regulation (SONAR) [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] uses text mining, statistical regression, rule-based
inference, uncertainty, and fuzzy matching. It mines for explicit and implicit relationships
among the entities and events, all of which form episodes or scenarios with specific
identifiers. Yamanishi et al. [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] reduce the problem of change point detection in time
series into that of outlier detection from time series of moving-averaged scores. Ge et
al. [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] extend hidden semi Markov model for change detection. Both these solutions
are applicable to different data distributions using different regression functions;
however, they are not scalable to large size datasets due to their time complexity.
      </p>
    </sec>
    <sec id="sec-5">
      <title>5 Peer Group Analysis</title>
      <sec id="sec-5-1">
        <title>5.1 Overview</title>
        <p>The following processes are involved in PGA (fig. 1).</p>
        <sec id="sec-5-1-1">
          <title>Data</title>
          <p>Modeling</p>
        </sec>
        <sec id="sec-5-1-2">
          <title>Statistical analysis such as mean</title>
        </sec>
        <sec id="sec-5-1-3">
          <title>Similar objects (peer group) identification</title>
        </sec>
        <sec id="sec-5-1-4">
          <title>Comparing target object with peer group objects</title>
        </sec>
        <sec id="sec-5-1-5">
          <title>Flagging transactions which deviate from peer groups</title>
          <p>Fig 1. Overview of PGA</p>
          <p>Peer group analysis (PGA) is a term that have been coined to describe the analysis
of the time evolution of a given object (the target) relative to other objects that have
been identified as initially similar to the target in some sense (the peer group).</p>
          <p>Since PGA finds anomalous trends in the data, it is reasonable to characterize such
data in balanced form by collating data under fixed time periods. For example, the
total quantity sold can be aggregated per week or the number of phone calls can be
counted per day.</p>
          <p>After the data modeling process, some statistical analyses are required. Mean or
variance can be appropriate. In our research we used weekly mean of stock
transactions.</p>
          <p>Then the most important task of PGA method is to identify peer groups for all the
target observations (objects). Member of peer groups are the most similar objects to
the target object. In order to make the definition of peer group precise, we must
decide how many objects, npeer, it contains from the complete set of objects. The
parameter npeer effectively controls the sensitivity of the peer group analysis. Of
course, if npeer is chosen to be too small then the behavior of the peer group may be
too sensitive to random errors and thus inaccurate. The length of time window for
calculating the peer group can be chosen based on the particular data set. We used 5
weeks for our experiments.</p>
          <p>Peer groups are summarized at each subsequent time point and the target object is
then compared with its peer group’s summary. Those accounts deviate from their peer
groups more substantially are flagged as outliers for further investigation. The
processes from the peer group identification to the account flagging are repeated as long as
the proper result is found.</p>
        </sec>
      </sec>
      <sec id="sec-5-2">
        <title>5.2 Significance of PGA</title>
        <p>The approach of PGA is different in that a profile is formed based on the behavior of
several similar users where current outlier detection techniques over time include
profiling for single user.</p>
        <p>A point may not be seen as unusual when compared with the whole set of points
but may display unusual properties when compared with its peer group. This is the
most significance feature of PGA.</p>
      </sec>
      <sec id="sec-5-3">
        <title>5.3 Definition of Peer Groups</title>
        <p>
          Based on [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ], Let us suppose that we have observations on N objects, where each
observation is a sequence of d values, represented by a vector, x i , of length d. The jth
value of the ith observation, x ij , occurs at a fixed time point t j .
        </p>
        <p>Let PG i (t j ) = {some subset of observations (≠x i ), which show behavior similar
to that of x i at time t j }. Then PG i (t j ) is the peer groups of object i, at time j.</p>
        <p>The parameter npeer describes the number of objects in the peer group and
effectively controls the sensitivity of the peer group analysis. The problem of finding a
good number of peers is akin to finding the correct number of neighbors in a
nearestneighbor analysis.</p>
      </sec>
      <sec id="sec-5-4">
        <title>5.4 Peer Group Statistics</title>
        <p>Let S ij be a statistic summarizing the behavior of the ith observations at time j. Once
we have found the peer group for the target observation x i we can calculate peer
group statistics, P ij . These will generally be summaries of the values of S ij for the
members of the peer group. The principle here is that the peer group initially provides
a local model, P i1 , for S i1 , thus characterizing the local behavior of x i at time t 1 ,
and will subsequently provide models, P ij , for S ij , at time t j , j&gt;1. If our target
observation, S ik , deviates ‘significantly’ from its peer group model P ik at time t k , then
we conclude that our target is no longer behaving like its peers at time t k . If the
departure is large enough, then the target observation will be flagged as worthy of
investigation.</p>
        <p>To measure the departure of the target observation from its peer group we calculate
its standardized distance from the peer group model; the example we use here is a
standardized distance from the centroid of the peer group based on a t-statistic. The
centroid value of the peer group is given by the equation 1:
(1)
(2)
(3)
where PG i (t 1 ) is the peer group calculated at time t 1 . The variance of the peer
group can be calculated by the equation 2:
where j  1, p  i .</p>
        <p>The square root of this can be used to standardize the difference between the target
S ij and the peer group summary P ij , yielding equation 3:</p>
        <p>P ij =</p>
        <p>1   S pj  ; j  1, p  i .</p>
        <p>npeer  pPGi t1  
V ij =
1 2</p>
        <p> S pj  Pij  .</p>
        <p>npeer  1 pPGi t1 
T ij = Sij  Pij  </p>
      </sec>
      <sec id="sec-5-5">
        <title>6.1 Experimental Data</title>
        <p>Our data set consists of 3 months real data from 6/01/2005 to 08/31/2005 for the daily
stock sell quantity and the number of transactions for each of 143 brokers. The total
number of transaction is 340,234. This data has been collected from the Dhaka Stock
Exchange (Bangladesh).</p>
        <p>Here we set, d = 14 weeks, N = 143. The length of time window, w = 5, but varied
npeer to take values npeer = 13 and npeer = 26. A sample of stock market data has
shown in table 2:</p>
      </sec>
      <sec id="sec-5-6">
        <title>6.2 Experimental Results</title>
        <p>For comparison purpose, we simulated PGA over stock transactions many times by
changing the number of peers. Here we have shown some of the results, which are
more interesting. The following plots illustrate the power of PGA to detect local
anomalies in the data. The vertical axis shows cumulative stock sold as weeks pass on
the horizontal axis. The sale quantity of the target observation is represented by thick
black line and the sale quantities of the peer group are represented by black dotted
lines. The graph of number of transactions has shown in the same manner.</p>
        <p>Fig 2. PGA Over Sell Quantity, Account # 132 npeer =13
Fig 3. PGA Over Sell Quantity, Account # 132 npeer = 26.</p>
        <p>Fig 4. PGA Over Sell Quantity, Account # 68 npeer = 13</p>
        <p>
          We have also measured the departure of the target observation from its peer group.
If the departure is large enough then the target observation will be flagged as worthy
of investigation. For this purpose we have calculated its standardized distance from
the peer group model. Table 3 shows the standardized distances from the centroid of
the peer group based on a t-statistic [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ].
Fig 2 shows account132 is flagged since it has the highest suspicious score in the 8th
week. Fig 3 shows account 132 where npeer is increased to 26. These simulations
were conducted according to the traditional PGA but with a slight change in
parameters npeer and the time window. The behavior of this account varied largely from its
peers almost in every week even though number of peers was increased. According to
the suspicious score calculated by t-statistics (Table 3), account 132 is the most
suspicious one. This is an outlier but it may not be a fraud case. Since the behavior of this
account is different from its peer group from the beginning, may be it is the general
nature of this particular broker. Even introduction of more npeer is not enough to
decide weather it is a fraud case or not. The account’s behavior is still far away from its
peers.
        </p>
        <p>Fig 4 shows account 68 is flagged since it has a clear sudden rise in 12th week
whereas most peers have very little sales in this week. This could be a possible fraud
case since the behavior of this account was quite similar to its peer group for all the
weeks except the sudden rise in the 12th week. Fig 5 shows account 68 where npeer is
26. Here we got very interesting findings. The behavior of this account also has not
been affected by the increase of npeer but it makes the account more suspicious.
According to our proposed method in section 3, we extended the investigation for the
suspicious accounts in fig 6. We included another attribute into the outlier mining
process. The main idea is to evaluate more information before flagging the fraudulent
account, where traditional PGA considers only one attribute to flag an account.</p>
        <p>We considered the number of transactions as well as trade volume as another
indicator of stock fraud. Fig 6 shows PGA over number of stock transactions for the same
account 68. It is obvious from the figure 6 that the number of transactions has
suspiciously increased in the 12th week.</p>
        <p>So, from fig 4, 5 and 6 we can do a comparative analysis about account 68. The
discovered knowledge here is:
 Account 68 has sudden rise for both the sale quantity and the number of
transactions in 12th week.
 The behavior of this account was similar with its peer group’s sale
quantity and number of transactions in other weeks.</p>
        <p>Now we can flag the account as an outlier or possible fraud case more confidently,
because both the observations have same findings. The process of calculating the peer
groups and t-scores can be run every minute in a real-time manner. Thus the outlier
mining process becomes more effective than the one-attribute observation process.</p>
        <p>Here we have demonstrated the results with the suitable npeer for our data set. In
practical application, the flagged accounts will simply be noted as meriting more
detailed examination. Using over 340,234 transactions gives an indicator of the
performance of PGA on large data sets.
8</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Conclusions and Future Work</title>
      <p>In this paper, we tried to mention the necessity of stock market fraud detection since
the area has lack of proper researches. We have demonstrated the experimental results
of PGA tool in an unsupervised problem over real stock market data sets with
continuous values over regular time intervals. The visual evidences have been shown
through graphical plots that peer group analysis can be useful in detecting
observations that deviate from their peers. We also applied t-statistics to find the deviations
effectively.</p>
      <p>In future, we aim to investigate for weather PGA can identify labeled fraudulent
objects or not from a real fraud data set. To make the stock fraud detection more
effective we will mine the following cases of possible outliers:
 To identify stock IDs and buyer IDs in case of trade volume or trade
quantity increases suspiciously.
 To identify stock IDs with sudden raise or fall in price or having same
buyer and seller.</p>
      <p>We have intention to integrate some other effective methods with PGA. We will
also apply our strategy on other more applications, such as banking fraud detection,
network intrusion detection etc.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Barnett</surname>
            <given-names>V.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Lewis</surname>
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Outliers in Statistical Data</article-title>
          . John Wiley, New York, 3rd edition, (
          <year>1994</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2. Johnson R.:
          <article-title>Applied Multivariate Statistical Analysis</article-title>
          .
          <source>Prentice Hall</source>
          , (
          <year>1992</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Victoria</surname>
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Hodge</surname>
          </string-name>
          , Jim Austin:
          <article-title>A Survey of Outlier Detection Methodologies</article-title>
          . Artif. Intell. (
          <year>2004</year>
          ).
          <volume>22</volume>
          (
          <issue>2</issue>
          ):
          <fpage>85</fpage>
          -
          <lpage>126</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Knorr</surname>
            <given-names>E.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Ng</surname>
            <given-names>R.</given-names>
          </string-name>
          :
          <article-title>Algorithms for Mining Distance-Based Outliers in Large Datasets</article-title>
          .
          <source>In Proceedings of 24th VLDB Conference</source>
          , (
          <year>1998</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Ruts</surname>
            <given-names>I.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Rousseeuw P.</surname>
          </string-name>
          :
          <article-title>Computing Depth Contours of Bivariate Point Clouds</article-title>
          .
          <source>Computational Statistics and Data Analysis</source>
          , (
          <year>1996</year>
          )
          <volume>23</volume>
          :
          <fpage>153</fpage>
          -
          <lpage>168</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Johnson</surname>
            <given-names>T.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Kwok</surname>
            <given-names>I.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Ng</surname>
            <given-names>R.</given-names>
          </string-name>
          :
          <article-title>Fast Computation of 2-Dimensional Depth Contours</article-title>
          .
          <source>In Proceedings of the Fourth International Conference on Knowledge Discovery and Data Mining</source>
          (
          <year>1998</year>
          ),
          <fpage>224</fpage>
          -
          <lpage>228</lpage>
          . AAAI Press.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Bolton</surname>
            ,
            <given-names>R. J.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Hand D</surname>
          </string-name>
          . J.:
          <article-title>Unsupervised Profiling Methods for Fraud Detection, Credit Scoring and Credit Control VII, Edinburgh</article-title>
          , UK (
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Qu</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vetter</surname>
            <given-names>B. M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Narayan</surname>
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wu</surname>
            <given-names>S. F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hou</surname>
            <given-names>Y. F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gong</surname>
            <given-names>F.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Sargor</surname>
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Statistical Anomaly Detection for Link-State Routing Protocols</article-title>
          .
          <source>In Proceedings of the Sixth International Conference on Network Protocols</source>
          (
          <year>1998</year>
          ),
          <fpage>62</fpage>
          -
          <lpage>70</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Lane</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Brodley</surname>
            <given-names>C. E.</given-names>
          </string-name>
          :
          <article-title>Temporal Sequence Learning and Anomaly Detection</article-title>
          .
          <source>In Proceedings of the 5th ACM Conference on Communications Security</source>
          . (
          <year>1998</year>
          ),
          <fpage>150</fpage>
          -
          <lpage>158</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Forrest</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hofmeyr</surname>
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Somayaji</surname>
            <given-names>A.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Longstaff</surname>
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>A sense of self for unix processes</article-title>
          .
          <source>In Proceedings of the IEEE Symposium on Security and Privacy</source>
          , (
          <year>1996</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Kosoresow</surname>
            ,
            <given-names>A. P.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Hofmeyr S</surname>
          </string-name>
          . A.:
          <article-title>Intrusion Detection via System Call Traces</article-title>
          . IEEE Software,
          <article-title>(</article-title>
          <year>1997</year>
          ),
          <volume>14</volume>
          (
          <issue>5</issue>
          ),
          <fpage>24</fpage>
          -
          <lpage>42</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Maes</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tuyls</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vanschoenwinkel</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Manderick</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Credit Card Fraud Detection using Bayesian and Neural Networks</article-title>
          .
          <source>In Proceedings of the 1st International NAISO Congress on Neuro Fuzzy Technologies</source>
          (
          <year>2002</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Goldberg</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kirkland</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shyr</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Thakker</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>The NASD Securities Observation, News Analysis &amp; Regulation System (SONAR)</article-title>
          .
          <source>In Proceedings of IAAI</source>
          , (
          <year>2003</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Yamanishi</surname>
            <given-names>K.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Takeuchi</surname>
            <given-names>J.:</given-names>
          </string-name>
          <article-title>A unifying framework for detecting outliers and change points from non-stationary time series data</article-title>
          .
          <source>In Proceedings of the eighth ACM SIGKDD international conference on Knowledge discovery and data mining</source>
          , (
          <year>2002</year>
          ),
          <fpage>676</fpage>
          -
          <lpage>681</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Ge</surname>
            <given-names>X.</given-names>
          </string-name>
          :
          <article-title>Segmental semi-markov models and applications to sequence analysis</article-title>
          ,
          <source>PhD thesis</source>
          ,
          <string-name>
            <surname>Chair-Padhraic</surname>
            <given-names>Smyth</given-names>
          </string-name>
          , (
          <year>2002</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Hand</surname>
            <given-names>D.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mannila</surname>
            <given-names>H.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Smyth</surname>
            <given-names>P</given-names>
          </string-name>
          :
          <article-title>Principles of Data Mining</article-title>
          , MIT Press (
          <year>2001</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>