<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Case-Base Maintenance: A Streaming Approach</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Yang Zhang</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Su Zhang</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>David Leake</string-name>
          <email>leake@cs.indiana.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>School of Informatics and Computing, Indiana University</institution>
          ,
          <addr-line>Bloomington, IN 47405</addr-line>
          ,
          <country country="US">U.S.A</country>
        </aff>
      </contrib-group>
      <fpage>222</fpage>
      <lpage>231</lpage>
      <abstract>
        <p>Case Base maintenance can be crucial to CBR system performance. A central current of case base maintenance research focuses on competence-based deletion. Traditionally, deletion is done periodically, pausing CBR system processing to examining the entire case base. However, for streaming data, as often arises in big data contexts, such an approach may be expensive or infeasible. To address this problem, this paper proposes that CBR maintenance can draw on advances from data discovery research. In particular, it presents a case study applying the recent Sieve-Streaming algorithm [2] to enable continuous streaming CBR maintenance to reduce demands on case storage and provide e cient continuous maintenance. The paper presents a preliminary evaluation of this method on the Travel Agent Case Base, and compares it to traditional methods. The experiments are encouraging for the practicality and bene ts of the approach for scaleup of case-base maintenance in settings with large-scale data streams.</p>
      </abstract>
      <kwd-group>
        <kwd>Case-Base Maintenance</kwd>
        <kwd>Case Deletion</kwd>
        <kwd>Streaming Algorithm</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Case-based Reasoning(CBR) is the process of reasoning by adapting prior
relevant cases to solve new problems (e.g., [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]). As CBR systems are used over
long time frames, system knowledge must be maintained over time. For
maintenance of the case base, a common question is how maintenance strategies can
adjust knowledge to balance system e ciency and solution quality [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. This has
often focused on how to increase retrieval e ciency, by decreasing case base size,
while maintaining system competence. Extensive CBR research has addressed
this through methods for competence-based deletion, building on the seminal
work of Smyth and Keane [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. Such methods have been shown to provide good
results for compression while retaining system accuracy.
      </p>
      <p>
        As the CBR community addresses issues for growing data sets, issues arise
making such models more di cult to apply. Competence-based deletion
approaches often use greedy methods to selecting a subset of cases to retain from a
case base to maximize competence, starting from a \complete" case base whose
competence they aim to retain; processing is done in a maintenance step to which
it is assumed that considerable resources can be devoted, pausing the CBR cycle.
Such policies can be characterized as synchronic, processing a snapshot of the
case base, and periodic [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] with comparatively large periodicity.
      </p>
      <p>However, CBR applications may need to be elded with small sets of seed
cases, with cases acquired from processing over time, and may be applied in
domains that change over time. For such systems, a \complete" case base may
never exist. Even if a complete case base might exist in principle, such a case
base might be prohibitively large. Large data sets are acquired on a daily basis
by streaming systems. For example, in the context of e-commerce, on \Cyber
Monday" of 2016, Amazon U.S. reported that it processed 23 million orders.
Building a comparatively complete case base from large-scale steaming data
could require enormous storage and expensive processing to compress the case
base. Even if storage were available for all cases in a large scale system, it might
be undesirable in real-time CBR systems to interrupt system processing for
frequent compression of the case base, if cases were received at a high rate.</p>
      <p>
        The general problem of dealing with large scale streaming data is of course
not new to CBR. In fact, the question of summarizing large-scale streaming data
is an active research topic for the knowledge discovery community. This raises
the question of whether approaches from that community provide solutions that
the CBR community can exploit to improve case base maintenance performance.
This paper presents a case study of the application of a recent method for
massive data summarization, Sieve-Streaming [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], to case-base maintenance. Sieve
streaming is a knowledge discovery method for selecting representative elements
of large data sets in a single pass through data, with xed memory requirements,
with performance guarantees to provide guarantees on level of approximation of
an optimal solution. The approach is been tested successfully for applications
such as on-the- y clustering.
      </p>
      <p>This paper presents a preliminary exploration of the use of Sieve-Streaming
for continuous incremental case base maintenance of cases from a case stream,
without access to the full case base. It describes this approach and evaluates it
for the Travel Agents Case Base, comparing its performance in terms of case base
size, competence and retrieval quality of case base with two baseline methods.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>
        Case-base maintenance has been the subject of extensive study, summarized in
the literature (e.g. [
        <xref ref-type="bibr" rid="ref12 ref7">7, 12</xref>
        ]). Most relevant to this paper are the two approaches we
will take as baselines. Condensed Nearest Neighbor (CNN) [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] is a classic method
for compaction of data sets to be solved by Nearest Neighbor algorithms, but
CNN is order-dependent and requires multiple loops over data set to guarantee
the consistency.
      </p>
      <p>
        Smyth and McKenna[
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] developed a competence-based model of CBM and
de ned the concept of coverage, reachability and relative coverage which modeled
local competence contributions of cases interact as well as global competence
properties of a case base. This gave rise to several competence-guided editing
solutions, called footprinting techniques, such as COV-FP, RFC-FP, RC-FP, etc,
and the footprint deletion(FD) algorithm.
3
      </p>
    </sec>
    <sec id="sec-3">
      <title>Adapting the Sieve-Streaming Algorithm for Case Base</title>
    </sec>
    <sec id="sec-4">
      <title>Maintenance</title>
      <p>3.1</p>
      <sec id="sec-4-1">
        <title>The Sieve-Streaming Algorithm</title>
        <p>
          The Sieve-Streaming algorithm is a method to \summarize a massive data set
\on the y" [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] by selecting a high utility subset. The utility of the subset re ects
the utilities of its members, according to a utility function whose values depend
not only on the member being evaluated but on the other elements of the subset.
The task addressed by Sieve-Streaming maps naturally to the task of selecting
a high competence subset of the case base, with overall utility corresponding to
overall competence, and the value of including particular cases depending on the
other cases in the case base.
        </p>
        <p>Sieve-Streaming addresses the problem of selecting elements to retain for
very large data sets, for which storing the entire set is not possible or
practical. If the optimal solution (for the case-base maintenance problem, the optimal
competence) were known, it would be possible to strategically retain only those
elements in the stream that provide su cient gain towards the optimal solution.
As the optimal solution will generally not be known, Sieve-Streaming
approximates knowledge of the optimum by using a collection of estimations re ned
as new elements are seen. Each particular estimation corresponds to a set of
elements, called a sieve. The number of sieves used, and the particular sieves
which are retained, changed dynamically as the data stream is processed and
more information becomes available. Incoming data is processed by each sieve,
and the algorithm's result is the set contained in the sieve with maximal value.
This approach enables calculating a reasonable solution from a subset of the
values. In a CBR context, we can consider each sieve be a candidate case base,
built in parallel, with the most successful returned by the algorithm.</p>
        <p>
          The Sieve-Streaming algorithm assumes that a maximal data set (case base)
size, k, is prede ned. It tracks m, the maximum utility value of all individual
examples seen, using that to estimate an upper bound on the marginal utility
contribution of each added instance. As m changes, the algorithm adjusts the set
of candidate thresholds determining the marginal utility to retain an example.
For each input element ei, the algorithm calculates the marginal gain f (eijSv)
of adding the element ei to sieve Sv, given the a prede ned utility function
f (S). If the marginal gain is more signi cant than the marginal value threshold,
v
calculated as 2 f(Sv) , and the maximum sieve size k has not yet been reached,
k jSvj
the data point will be added to the sieve; otherwise, it will be discarded. Finally,
the algorithm outputs the elements in the best sieve as output. Full details,
including performance guarantees, are provided by Badanidiyuru et al. [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ].
3.2
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>Adding Sieve-Streaming Maintenance to the CBR Cycle</title>
        <p>A motivation for incorporating Sieve-Streaming into CBR is to enable
efcient ongoing competence-based case base updates at every execution of the
CBR cycle. Sieve-Streaming can be applied to case-base maintenance as follows.
First, a competence-based utility function is de ned to calculate marginal gain
for each case. The following subsection presents a simple domain-independent
sample choice, but domain-speci c utility functions could be de ned as well.</p>
        <p>Given an input problem, the system retrieves the most similar case in its
case base. If the problem solved by the retrieved case is not identical to the new
problem, the new problem is ltered by Sieve-Streaming to determine whether
to add it to the case base. The case's marginal gain is compared to that of sieves
whose capacity constraint has not yet been reached. If its marginal gain is larger
than this sieve's marginal value threshold, the case is added to this sieve. If
case's marginal gain value smaller than every sieve's marginal value threshold,
the case is not added and processing continues with the next problem.</p>
        <p>This approach enables the system to maintain a dynamic case base within
desired size limits, with coverage satisfying the sieve streaming algorithm's
guarantees, with ongoing rather than periodic maintenance.
3.3</p>
      </sec>
      <sec id="sec-4-3">
        <title>A Sample Competence-based Utility Function</title>
        <p>
          In order to apply Sieve-Streaming, a utility function must be de ned, based
on the system's similarity metric, to develop a submodular objective function f
re ecting competence contributions. The goal for case base maintenance is to nd
a compact representative subset, which has characteristics similar to exemplar
selection. Thus, we model our application on the example of Badanidiyuru et al.
for exemplar selection, with K-medoid loss function as the utility function.
K-medoid Loss Function The K-medoid method selects elements by
minimizing the sum of pairwise distances[
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]. Following Badanidiyuru et al. [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ], here
distance(e; v) is used to encode distances between exemplar and element. This
yields Badanidiyuru et al.'s loss function:
L(S) =: 1 X min distance(e; v)
jV j e2V v2S
(1)
(2)
        </p>
        <p>
          For each element in V , the algorithm calculates distance between it and any
element in given set S and selects the minimum distance among the results.
Then it calculates the average of this minimum distance and used as the loss
of set S. We want to minimize this loss. Following Krause and Gomes [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ], L(S)
can be transformed into a monotone submodular way with an auxiliary element
e0, resulting in the following de nition from Badanidiyuru et al.:
f (S) =: L(fe0g)
        </p>
        <p>L(S [ fe0g)</p>
        <p>Here, f (S) is always greater than 0 no matter what e0 is chosen. Then
minimizing L is equal to maximizing f . Because we also use similarity to measure the
competence, we could use this f (S) as the utility function in the Sieve-Streaming
algorithm.</p>
        <p>The calculation of loss relies on V {the entire data set{which could never be
accessed during processing. However, we can address this in practice by using a
sample of cases from the initial case base as an evaluation set.
4</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Evaluation</title>
      <p>The evaluation of streaming case retention explored four questions, each
examining behavior as a function of the number of problems processed:
1. How does maintenance processing time vary for streaming case retention
compared to baselines?
2. How does case base size vary for streaming case retention?
3. How does case base competence vary, and how does it compare to that with
baseline methods?
4. How does accumulated retrieval distance vary, and how does it compare to
that with baseline methods?
4.1</p>
      <sec id="sec-5-1">
        <title>Experimental Design</title>
        <p>Test case base: Tests were performed using the Travel Agents Case Base,1
commonly used as a benchmark within the CBR community. This case base contains
1470 cases, each with a case identi er plus 9 features: Journey Code, Holiday
Type, Price, Number of Persons, Region, Transportation, Duration, Season,
Accommodation, Hotel. There are 6 categorical features and 3 numerical features
1 http://cbrwiki.fdi.ucm.es/mediawiki/index.php/Case_Bases
1. Sampling parameters
(a) n: Size of input stream.
(b) evaluation size: Size of evaluation set. This sampled from same domain but
was exclusive of the input stream.</p>
        <p>(c) init size: Size of initial case base (randomly selected)
2. Sieve-Streaming parameters
(a) k : Restricted maximum size of sieves' capacity, also the size of case base.
(b) : Scale of threshold set O; 2 [0; 1].
3. CNN-FP/RC-FP parameters
(a) max : Maintenance threshold. If Case-Base size increased to this value,
CNN</p>
        <p>
          FP or RC-FP maintenance method will be triggered.
(b) min: Maintenance threshold. After maintenance, Case Base size should always
no larger than this value.
(c) solve threshold : Cases within this threshold of each other are considered to
\solve" each other
left in each case. Similarity of categorical data was measured with Hu man Code
Similarity based on Hu man coding [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]; this was used to translate categorical
data to binary strings then converted to decimal values, enabling measuring
similarity between cases simply by calculating Euclidean distance.
Competence Calculation: Smyth and McKenna's Relative Coverage [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ] is an
in uential method for calculating competence contributions. In general, the
calculation of Relative Coverage relies on determination of the the Coverage Set
and Reachability Set for a case. However, in travel domain, our CBR system is a
simple recommendation system which accepts trip characteristics and retrieves
a similar case to return. This simpli es the calculation of Relative Coverage; we
simply use similarity measurements to represent competence.
4.2
        </p>
      </sec>
      <sec id="sec-5-2">
        <title>Test Runs and Parameter Settings</title>
        <p>Experimental runs were conducted on four input streams. These were
generated by randomly selecting an ordered set of 1000 of the 1470 cases in Travel
Agent Case Base, and taking pre xes of lengths 100, 200, 500, and 1000 cases
respectively.</p>
        <p>The tests assessed the performance of Sieve-Streaming Algorithm along with
CNN-FP and RC-FP as benchmark methods. Table 1 describes the parameters of
the algorithms; Table 2 reports their settings for all runs. Note that for di erent
size streams, di erent parameter settings were used for initial case base sizes
and the minimum and maximum values of the case base size for non-streaming
methods (always varying from 5% to 10% of the entire case base).
Sampling</p>
        <p>Sieve</p>
        <p>Streaming
CNN-FP/RC-FP
Question 1: Time Comparison: Table 3 shows that in our tests, CNN-FP
and Sieve-Streaming had similar time performance, with Sieve-Streaming
more expensive. This was a surprising result. However, timings may depend
on implementation factors and no attempt was made to optimize the
SieveStreaming algorithm; more investigation is needed. Both CNN-FP and
SieveStreaming ran much faster than RC-FP, and RC-FPs growth rate was much
higher than for CNN-FP and Sieve-Streaming.
Question 2: Case Base Size: Figure 2 shows changing case base size during
processing. CNN-FP and RC-FP size uctuates, growing from the minimum
to the limit, while Sieve-Streaming tends to maintain a smaller and more
stable Case-Base size than CNN and RC.</p>
        <p>Question 3: Competence: Figure 3 shows competence with the three
methods. Competence is assessed using a randomly selected evaluation set, drawn
from cases outside the case stream (the same set is used for all streams).
Competence is calculated with the evaluation set based on the k-medoid
utility function. CNN-FP and RC-FP show better initial performance for all
streams, with some uctuation as case base size changes. The competence of
Sieve-Streaming increases gradually, nally surpassing CNN-FP and RC-FP.
(a) n = 100
Question 4: Retrieval Quality: We use accumulated distances between each
incoming case and correspond retrieval case to measure retrieval accuracy
to re ect the e ciency of maintenance strategies. Figure 4 shows that
RCFP always has the lowest accumulated retrieval distances. CNN has a lower
accumulated retrieval distances over Sieve-Streaming for n=100 and n=200,
however, when n=500, accuracy of Sieve-Streaming is better.
5</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Conclusion</title>
      <p>Knowledge discovery methods provide an important potential resource for CBR.
This paper has explored the use of Sieve-Streaming for on-the- y, incremental
competence-based case deletion. The results suggest that it provides high
quality case base compression, trading o some accuracy loss compared to the gold
standard method RC-FP for much improved e ciency. We noted that the
performance of Sieve-Streaming improved when the size of the case stream increased.
A surprising result was that Sieve-Streaming required more processing time than
CNN-FP. This may be an artifact of implementation, especially given that the
implementations were not optimized. We are currently re-implementing all
methods to enable more uniform comparison. The superior competence results suggest
that the use of Sieve-Streaming compared to CNN-FP would still be justi ed,
(a) n = 100
but additional examination is needed to better understand the processing times
required for all methods.</p>
      <p>The results presented here are preliminary in that they report a limited set
of trials, on a small case base, without a systematic e ort to tune parameters. In
future work, we plan to test the strategies with multiple large scale data streams
and to examine further how the approach is a ected by changing parameter
settings.
(a) n = 100</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Aamodt</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Plaza</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          :
          <article-title>Case-based reasoning: Foundational issues, methodological variations, and system approaches</article-title>
          .
          <source>AI Communications</source>
          <volume>7</volume>
          (
          <issue>1</issue>
          ),
          <volume>39</volume>
          {
          <fpage>52</fpage>
          (
          <year>1994</year>
          ), http://www.iiia.csic.es/People/enric/AICom.pdf
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Badanidiyuru</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mirzasoleiman</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Karbasi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Krause</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Streaming submodular maximization: Massive data summarization on the y</article-title>
          .
          <source>In: Proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining</source>
          . pp.
          <volume>671</volume>
          {
          <fpage>680</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Hart</surname>
            ,
            <given-names>P.E.</given-names>
          </string-name>
          :
          <article-title>The condensed nearest neighbor rule</article-title>
          .
          <source>IEEE Transactions on Information Theory</source>
          <volume>14</volume>
          ,
          <volume>515</volume>
          {
          <fpage>516</fpage>
          (
          <year>1968</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Kaufman</surname>
          </string-name>
          , L.,
          <string-name>
            <surname>Rousseeuw</surname>
            ,
            <given-names>P.J.:</given-names>
          </string-name>
          <article-title>Finding groups in data: an introduction to cluster analysis</article-title>
          , vol.
          <volume>344</volume>
          . John Wiley &amp; Sons (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Krause</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gomes</surname>
          </string-name>
          , R.G.:
          <article-title>Budgeted nonparametric learning from data streams</article-title>
          .
          <source>In: Proceedings of the 27th International Conference on Machine Learning (ICML-10)</source>
          . pp.
          <volume>391</volume>
          {
          <issue>398</issue>
          (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Leake</surname>
            ,
            <given-names>D.B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Smyth</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          , Wilson,
          <string-name>
            <given-names>D.C.</given-names>
            ,
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <surname>Q.</surname>
          </string-name>
          :
          <article-title>Introduction to the special issue on maintaining case-based reasoning systems</article-title>
          .
          <source>Computational Intelligence</source>
          <volume>17</volume>
          (
          <issue>2</issue>
          ),
          <volume>193</volume>
          {
          <fpage>195</fpage>
          (
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7. Lopez de Mantaras, R.,
          <string-name>
            <surname>McSherry</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bridge</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Leake</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Smyth</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Craw</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Faltings</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Maher</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cox</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Forbus</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Keane</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Aamodt</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Watson</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Retrieval, reuse, revision, and retention in CBR</article-title>
          .
          <source>Knowledge Engineering Review</source>
          <volume>20</volume>
          (
          <issue>3</issue>
          ) (
          <year>2005</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>McKenna</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Smyth</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Building compact competent case-bases</article-title>
          .
          <source>In: Case-based reasoning research and development</source>
          , pp.
          <volume>329</volume>
          {
          <fpage>342</fpage>
          . Springer (
          <year>1999</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Prasad</surname>
            ,
            <given-names>B.V.V.S.:</given-names>
          </string-name>
          <article-title>A novel distance similarity measure on learning techniques and comparison with image processing</article-title>
          .
          <source>International Journal of Engineering Research and Development</source>
          <volume>2</volume>
          (
          <issue>8</issue>
          ),
          <volume>29</volume>
          {
          <fpage>32</fpage>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Smyth</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Keane</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Remembering to forget: A competence-preserving case deletion policy for case-based reasoning systems</article-title>
          .
          <source>In: Proceedings of the Thirteenth International Joint Conference on Arti cial Intelligence</source>
          . pp.
          <volume>377</volume>
          {
          <fpage>382</fpage>
          . Morgan Kaufmann, San Mateo (
          <year>1995</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Smyth</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>McKenna</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          :
          <article-title>Competence models and the maintenance problem</article-title>
          .
          <source>Computational Intelligence</source>
          <volume>17</volume>
          (
          <issue>2</issue>
          ),
          <volume>235</volume>
          {
          <fpage>249</fpage>
          (
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12. Wilson,
          <string-name>
            <given-names>D.</given-names>
            ,
            <surname>Leake</surname>
          </string-name>
          ,
          <string-name>
            <surname>D.</surname>
          </string-name>
          :
          <article-title>Maintaining case-based reasoners: Dimensions and directions</article-title>
          .
          <source>Computational Intelligence</source>
          <volume>17</volume>
          (
          <issue>2</issue>
          ),
          <volume>196</volume>
          {
          <fpage>213</fpage>
          (
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>