<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Learning classification taxonomies from a classification knowledge based system</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Hendra</forename><surname>Suryanto</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Paul</forename><surname>Compton</surname></persName>
						</author>
						<title level="a" type="main">Learning classification taxonomies from a classification knowledge based system</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">90D524128871A58FEA58C5039A6A4AC0</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T12:26+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"> Knowledge-based systems (KBS)  <p>are not necessarily based on well-defined ontologies. In particular it is possible to build KBS for classification problems, where there is little constraint on how classes are organised and a class is expressed by the expert as a free text conclusion to a rule. This paper investigates how relations between such 'classes' may be discovered from existing knowledge bases, then investigates how to construct a model of these classes (an ontology) based on user-selected patterns in the class relations. We have applied our approach to KBS built with Ripple Down Rules (RDR) [1] RDR is a knowledge acquisition and knowledge maintenance methodology, which allows KBS to be built very rapidly and simply, but does not require a strong ontology. Our experimental results are based on a large real-world medical RDR KBS. The motivation for our work is to allow an ontology in a KBS to 'emerge' during development, rather than requiring the ontology to be established prior to the development of the KBS.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Most knowledge acquisition methodologies first build a model of domain knowledge, before using this to build a particular problem solver. e.g KADS and CommonKADS <ref type="bibr" target="#b1">[2]</ref>, Protege2000 <ref type="bibr" target="#b2">[3]</ref>. Although this approach facilitates reuse it does not overcome the knowledge acquisition and maintenance bottleneck and these problems are present both in the development of the ontology and consequent problem solver.</p><p>The RDR approach starts knowledge acquisition (KA) to build the problem solver immediately without any modelling apart from a simple attribute-value data representation <ref type="bibr" target="#b3">[4]</ref>. Even the attribute-value representation can be developed while KA is in progress. The focus of the approach is to make the addition of each incrementally added piece of knowledge as simple and as reliable as possible. Although this approach facilitates KA and maintenance <ref type="bibr" target="#b4">[5]</ref>, it does not facilitate re-use, because of the lack of an ontology.</p><p>In this learning problem, we are dealing with rules rather than raw cases; the relevant attributes have already been Artificial Intelligence Laboratory, School of Computer Science and Engineering, University of New South Wales, Sydney 2052, Australia, email: {hendras, compton}@cse.unsw.edu.au. identified and extracted. Our aim is to discover the appropriate ontology given that the relevant attributes (and values) are already well-identified. A second aspect of the problem is that in a real-world system, attributes are multivalued rather than boolean. In rules, conditions can then subsume each other, be disjoint, etc. For example age &gt;10 subsumes age 50, whereas age &gt;40 and age &lt;10 are clearly disjoint.</p><p>Our method needs to not only combine information about classes from across the knowledge base, but to address, the way in which conditions based on multivalued attributes interrelate. Figure <ref type="figure">.</ref> 1 shows some rules for the class Satisfactory lipid profile previous raised LDL noted. In the first rule there is a condition Max(LDL) &gt; 3.4 and in the second rule there is a condition Max(LDL) is HIGH), where HIGH is a range between 2 real number. Finally, the problem is compounded, by the way in which the expert adds conclusions.</p><p>When the Multiple Classification RDR (MCRDR) KB makes an error the task of the expert is to specify the correct conclusion and identify the attributes and values that justify this conclusion. In adding the conclusion, the expert can select from a list of pre-existing conclusions organised into broad categories, but can also simply type in a new conclusion. In medical pathology result interpretation, the evaluation domain here, the conclusions added by the pathologist may provide advice to the referring clinician: on patient diagnosis, management, how treatment is progressing, whether the tests ordered were appropriate, what tests might still be necessary or any combination of the above. It is quite clear to both the expert and the receiver of the advice what information is being provided in the free text interpretation, but these interpretations are a long way from well the defined classes of a formal ontology. A task analysis would assess this domain as a classification problem, but this does not imply well defined classes. Hence the problem is not only that disjuncts for a class (separate rule paths ) may be scattered across the KB, but that the same class may be represented by different text strings. Such text strings may cover a combination of different classes. Some examples are given in Table <ref type="table">1</ref>. The question arises of whether it would be better to start with a more well developed ontology, as suggested by most KA researchers. However in practice RDR systems allow experts to very rapidly and easily build large knowledge bases <ref type="bibr" target="#b4">[5]</ref> and recent commercial RDR systems have confirmed these advantages outside the research environment. (Pacific Knowledge Systems (PKS), personal communication) The aim of the present work then is to preserve the ease and speed of development provided by RDR systems, but overcome their lack of an initial strong ontology, by discovering the ontologies implicit in these incrementally developed systems. This may give us the best of both worlds.</p><p>RDR exception structure provides a compact representation of knowledge <ref type="bibr" target="#b5">[6]</ref><ref type="bibr" target="#b6">[7]</ref><ref type="bibr" target="#b7">[8]</ref><ref type="bibr" target="#b8">[9]</ref><ref type="bibr" target="#b9">[10]</ref><ref type="bibr" target="#b10">[11]</ref><ref type="bibr" target="#b11">[12]</ref><ref type="bibr" target="#b12">[13]</ref>. <ref type="bibr">Gaines ([14]</ref>) has also generalized RDR to Exception Directed Acyclic Graphs (EDAGs). He argues EDAGs are more compact and readable than MCRDR because of a graph rather than a tree structure. Initial RDR development was concerned with classification tasks, first single and later multiple classification. RDR has since been extended to configuration <ref type="bibr" target="#b14">[15]</ref>, heuristic search <ref type="bibr" target="#b15">[16]</ref>, document retrieval <ref type="bibr" target="#b16">[17]</ref> and a more general RDR system for construction tasks has been proposed <ref type="bibr" target="#b17">[18]</ref>.</p><p>Figure <ref type="figure" target="#fig_2">2</ref> shows a simple example of the exception structure used in MCRDR. All pathways are evaluated. Evaluation on any pathway stops when a leaf node is reached or no child rule is satisfied. A conclusion is provided by the last satisfied rule in each refinement pathway. When an expert identifies that an erroneous conclusion has been given, they enter a new rule, which the system adds as a refinement.</p><p>The expert is assisted in providing an appropriate rule by the system requiring that cases that correctly satisfy the parent rule should not satisfy the child rule.</p><p>In this study we have used 5 different pathology knowledge bases provided by PKS ranging in size from 25 to 320 rules. We also have associated sets of case data ranging in number from 453 to 2218 cases. The largest PKS knowledge base is over 7000 rules, but this is the subject of other related studies.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Ontology learning overview</head><p>Firstly we discover class relations between rules. We consider three basic relations: subsumption/intersection, mutual-exclusivity and similarity. Secondly we specify some compound relations which appear interesting using these three basic relations. We then extract the instances of these compound relations or patterns and assemble them into a class model.</p><p>We define that class-A subsumes class-B with subsumption value 1.0, if we always have class-A when we have class-B, but not the other way around. We decide that class-A subsumes class-B from the conditions used in the rules for class-A and class-B. This syntactical subsumption needs to be evaluated by the expert -whether class-A semantically subsumes class-B as well. If the subsumption value is less than 1.0 we use the term intersection rather than subsumption. If class-A subsumes class-B with a 0 value, it means we do not have any information about the subsumption/intersection relation for class-A and class-B.</p><p>The mutual exclusivity measure of class-A and class-B is 1.0 if class-A and class-B never occur together. In RDR we evaluate this measure by checking whether there is a condition in a rule for class-A mutually exclusive with any condition in a rule for class-B (for example age &gt; 50 and age &lt; 30). In RDR knowledge bases parent rules are always mutually exclusive with child rule.</p><p>Class-A similar to class-B with measure 1.0 if both of the classes have exactly the same conditions. We use the term same rather than similar if the similarity is 1.0.</p><p>A class in MCRDR is the set of disjunct rule paths giving the same conclusion. A rule path consists of all conditions from all predecessor rules plus conditions of this particular node's rule. For example, (see figure <ref type="figure" target="#fig_2">2</ref> The central idea of the technique is to group all rules for each class and compute a quantitative measurement (from 0 to 1) for each relation (subsumption, mutual-exclusivity, similarity) between every pair of classes. We use this quantitative measure as an informal confidence measure as to whether these relations exist. The algorithm will be discussed in detail below, but when applied to the example in figure <ref type="figure" target="#fig_2">2</ref>  Boolean values are obviously inappropriate for subsumption /intersection, mutual-exclusivity, and similarity relations in real domains. We found that in the Iron knowledge base, there were only 16 subsumption relations with degree of confidence 1.0; 8 mutually-exclusive relations with degree of confidence 1.0 and no similarity relations with degree of confidence 1.0.</p><p>We refine the subsumption/intersection measure by not only considering the conditions in the rule path but also the ratio of number of cases handled by parent and child rules in a rulepath. This allows us to deal with the situation where a rule is a gross over-generalisation and the child rule is added as correction to deal with most of the cases the parent rule would fire on. We consider that there is little value to the subsumption relation in this circumstance. For example (see figure <ref type="figure" target="#fig_2">2</ref>) rulepath-2 has two cases, rulepath-3 has two cases, rulepath-4 has 3 cases and rulepath-6 has 3 cases. We therefore calculate the quality of rulepath-2 is 2/(2+2+3+3) = 0.2, quality of rulepath-4 is 3/(3+3) = 0.5, quality of rulepath-3 is 2/2 = 1.0 and quality of rulepath-6 is 3/3 = 1.0.</p><p>We do not need a rule quality measure if we use flat rule system rather than the refinements of a rule path, for example a flat rule for rule_2 is: class Go Swimming ¡ sky=SUNNY, not(ultraviolet= VERY HIGH), not(wave=LOW).</p><p>Flat rules for rule_3 and rule_6 are the same as their rule paths since the rule path extends to the leaves and so no negation of child conditions is required. Rule_1 has three children with one, one and two conditions, and therefore we get (1 x 1 x 2) flat rules. By converting this RDR knowledge base to flat rules we get an equivalent knowledge base, but this is generally not feasible with real-world knowledge bases. In the five knowledge bases considered here, some rules have many children with several conditions. There is one rule with 10 children and every child has 5 conditions which converts to (5 10 ) flat rules.</p><p>One of the advantages of learning from rules is that we can assume that irrelevant attributes have already been discarded. This is significant as in our application domain there are hundreds of attributes. Gaines <ref type="bibr" target="#b18">[19]</ref> argues that a rule in a knowledge base is worth many cases for learning. We adopt the same viewpoint and note that although there is research on combining KA and machine learning and using background knowledge in machine learning, there seems little research so far in learning from a KBS rather than from cases <ref type="bibr" target="#b19">[20]</ref>, <ref type="bibr" target="#b13">[14]</ref>.</p><p>The immediate precursor of this work <ref type="bibr" target="#b19">[20]</ref> applied formal concept analysis to ontology discovery in knowledge bases. This provided a useful way to explore concepts in a knowledge base, but because of the complexity of the conceptual lattice it was more useful to consider sub-sections of the lattice, selected by the user or by a simple nearest neighbour algorithm <ref type="bibr" target="#b20">[21]</ref>. The critical difference from the work here, is that in formal concept analysis the difference between different concepts is emphasised. Here we attempt to combine all the concepts that represent a class and consider relations between classes rather than between concepts.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">The class relations model</head><p>The class relations model shows the relations subsumption, mutual-exclusivity and similarity between classes and the degree of confidence that the particular relation exists <ref type="bibr" target="#b21">[22]</ref>. We note that the measures we derive are strictly heuristic. Other superior and perhaps more well founded measures may be possible. The results here represent simply a first attempt at carrying out this type of analysis. The second point to note is that these relations have to deal with non-boolean as well as boolean data.</p><p>Let X be a class in the MCRDR framework. {X0…Xm} is a set of rules which have class X as the conclusion. {Xi0 … Xin} is a set of conditions for rule Xi where i = 0...n, n is the number of distinct conditions in the rule path; m is number of rule paths for class X. In the MCRDR framework the class is given as a disjunction of rule paths <ref type="bibr" target="#b6">[7]</ref>.</p><formula xml:id="formula_0">Then m n class X = ∨ ( ∧ X ij ) i=0 j=0</formula><p>That is, Xij stand for an individual condition in a rule path for the class X. If X is a class and Y is also a class, we could define a similarity measure as follows:</p><formula xml:id="formula_1">Sim (Xij ,Yij) = 0 if Xij ,Yij are different Sim (Xij ,Yij) = 1 if Xij ,Yij are same If α is set of distinct attributes in rule path X i , β is set of distinct attributes in rule path Y i , n = | α ∪ β|, (X ij ,Y ij</formula><p>) are pair of same attributes, but they could be different conditions, (e.g. Age&gt;50, Age=60) then we can define:</p><formula xml:id="formula_2">n Σ Sim(X ij ,Y ij ) j=0 Similar(X i ,Y i ) = −−−−−−−−−−−− * Q( X i ) * Q(Y i ) | α ∪ β|</formula><p>where Q( X i ) = number of cases for X i / (number of cases for descendant X i + number of cases for X i ). Function Q measures the quality of a rule in RDR.</p><p>If the quality of a rule is close to 100%, it means that nearly all cases reaching this rule are processed by the its child rules; and the child rules are rare exceptions. On the other hand if the quality of a rule is 10%, it means 90% of cases that satisfy that rule are passed to its children. That is the rule is too general and can be regarded as not being a particularly good rule and so it should not be given as strong consideration in developing the relations in the system.</p><p>For example Similar(rulepath-2, rulepath-6)</p><formula xml:id="formula_3">1/3 * Q(rulepath-2) * Q(rulepath-3), Similar(rulepath-8, rulepath-9) = 1/2 * Q(rulepath-8) * Q(rulepath-9),</formula><p>Similar(rulepath-9, rulepath-10) = 2/3 * Q(rulepath-9) * Q(rulepath-10).</p><p>Function Similar() measures a similarity between 2 nodes (each node contains a rule path).  We propose that ClassSimilarity(X,Y) = (v1 + v2 + v3) / 3, where we choose the v such that all nodes are covered by at least one edge and the sum of v (eg. v1 + v2 + v3) is maximal. Note that v stands for the Similar() function. In later similar diagrams v stands for the Subsume() and MutualEx() measures. Eg. ClassSimilarity(Go swimming, Play chess) = ((1/3)+(1/2)+(2/3))/3=0.5. We assume here the quality of all rule paths are 1.0 to simplify the example.</p><p>We can define a subsumption measure as follows with Xij and Yij standing for individual conditions in rule paths for the relevant classes as above.</p><formula xml:id="formula_4">Sub (Xij ,Yij) = 0 if Xij does not subsume Yij Sub (Xij, Yij) = 1 if Xij subsumes or same Yij (for example A&gt;5 subsumes A&gt;10)</formula><p>If α is set of distinct attributes in rule X i , β is set of distinct attributes in rule Y i , then we can define:</p><formula xml:id="formula_5">n Σ Sub(X ij ,Y ij ) j=0 Subsume(X i ,Y i ) = −−−−−−−−−−− * Q( X i ) * Q(Y i ) |α ∪ β|</formula><p>where Q(Xi ) = number of cases for Xi / (number of cases for descendant Xi + number of cases for Xi ) Function Subsume() measures a degree of confidence that the first rule path subsumes the second rule path. For example subsume(rulepath-2,rulepath-4) = 2/2 * Q(rulepath-2) * Q(rulepath-4).</p><p>Conditions in rulepath-2 are sky=SUNNY; conditions in rulepath-4 are sky=SUNNY, wave=LOW; and |α ∪ β| = 2, that is α ∪ β = {sky, wave}.</p><p>Since rulepath-2 does not have the attribute wave, we consider rulepath-2 is more general than rulepath-4 with the attribute wave. Therefore there are 2 conditions in rulepath-2 which are same as or more general than rulepath-4. Similarly, subsume (rulepath-2, rulepath-5) =2/3 * Q(rulepath-2) * Q(rulepath-5)</p><p>Figure <ref type="figure">5</ref> illustrates how we find a subsumption measure between 2 classes. It shows Class X as a disjunction of nodes 1,2 and 3 and Class Y as a disjunction of nodes 4,5 (each node contains a rule path).</p><p>We compute ClassSubsume(X,Y) = (v1 + v3) / 2. We choose the v such that all nodes of class Y are covered by at least one edge and the sum of v (eg. v1 + v3) is maximal. E.g. classSubsume ( Go swimming, Swimming in the beach) = (1+0.667)/2. We assume the quality of all rule paths is 1.0 to simplify this example.</p><p>We can define a mutually exclusive measure as follows with Xij and Yij standing for individual conditions in rule paths for the relevant classes as above.</p><p>Mut (Xij ,Yij) = 0 if Xij and Yij are not mutually exclusive. Mut (Xij ,Yij) = 1 if Xij and Yij are mutually exclusive (for example A&gt;5 and A&lt;2) MutualEx(Xi, Yi) = 1 , if at least one of Mut(Xij ,Yij)=1 MutualEx(Xi, Yi) = 1 , if rule Xi and rule Yi are parent and child.</p><p>MutualEx(Xi ,Yi) = 0 , otherwise Function MutualEx() measures a degree of confidence that the first rule subsumes the second rule. Since the quality of the rule does not affect mutual exclusivity as much as it affects similarity and subsumption/intersection, we do not apply the quality measure to mutual exclusivity. For example MutualEx(rulepath-2, rulepath-10) = 1.0, Figure <ref type="figure" target="#fig_9">6</ref> suggests how we can find a mutual exclusivity measure between 2 classes. It shows that Class X is a disjunction of nodes 1,2 and 3 and Class Y is a disjunction of nodes 4 and 5. We compute ClassMutualEx(X,Y) = (v1 + v2 +v3 +v4 + v5 + v6) / 6. X and Y are mutually exclusive if and only if all nodes of X and Y are mutually exclusive with respect to each other (see Figure <ref type="figure" target="#fig_9">6</ref>). Therefore ClassMutualEx (Go swimming, Play chess) = 1/6, since Go Swimming has 2 rulepaths, Play chess has 3 rulepaths and all MutualEx() between those rulepaths are each 0.0, except for MutualEx(rulepath-2, rulepath-10) = 1.0. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Experimental results</head><p>Table <ref type="table">1</ref>. some examples of class relations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Class relations model</head><p>The results from the endocrinology knowledge are shown in the table 1. The results shown in each table are the class pairs with the highest similarity, subsumption, or mutual exclusivity measures. Only results with high values for these relations are shown</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Extracting patterns from the class relation graph</head><p>Since there are many classes (from 25 to 100), it is impossible to consider all possible pairs of relations between the classes.</p><p>We therefore extract specific patterns which seem likely to be components of a meaningful taxonomy.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Discussion</head><p>It is beyond the scope of this paper (and the authors) to provide a detailed medical analysis of the ontologies produced by these techniques; however, it is worth noting some lay observations.</p><p>The mutually-exclusive classes in Table <ref type="table">1</ref> seem reasonable. However, we also have some cases where very similar conclusions are identified as mutually exclusive. This occurs when experts make up rules that include different values for the same attribute. For example some mutually exclusive conclusions seem to give the same clinical advice but specifically refer to a patient being male or female. This may or may not be of clinical importance, but there is an obvious opportunity to have a further superclass.</p><p>The most interesting issues arise with the nature of subsumption. A superclass subclass relation may arise where one rule specifies a value for an attribute and another does not. For example a key factor in comments 6, 7, 23, 30, 33 and 36 is whether or not a TSH measurement should be ordered and whether or not primary hypothyroidism has been excluded. TSH results are important in the diagnosis of primary hypothyroidism. The generic comment suggesting a TSH measurement is given when there is no TSH result available. The comment also suggests the clinical cause of the high prolactin level remains unknown. When there is a TSH result available this provides some evidence to confirm or exclude one of the causes of the high prolactin. These relations appears to us as ontologically reasonable. However, the wording of the actual comments does not readily indicate such relationships. It would be interesting to know how the expert would react and how comments might be worded if this ontological information was available as rules were being developed.</p><p>A more general example of this pattern is the comment [0]:"patient has ovulated" which is at the top level of the  This subsumes a whole range of more specific comments related to other attributes. Again it would be interesting to see if this taxonomic information influenced the expert's wording.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.">Further work</head><p>At this stage we have only conducted a preliminary examination of some of the relations in the knowledge bases.</p><p>A detailed examination by domain experts will be required to establish the utility of this approach. This examination may suggest that other types of measure than those suggested here may be useful and paterns other than those used in Fig 7 <ref type="figure">may</ref> be of interest for browsing the relations. At this stage we make no claim about the particular heuristic measures we have used, except that broadly, measures of this kind seem useful in discovering and exploring implicit ontologies.</p><p>We are also investigating the possibility of learning interesting patterns from a coloured graph of class relations. The frequency of isomorphic sub-graphs may be interesting.</p><p>If the graph is large, then we could scale up the algorithm by applying data mining technique (e.g the apriori algorithm <ref type="bibr" target="#b23">[24]</ref>)</p><p>Finally, the present technique only considers the conditions in rule paths and proportion of cases handled by rules in determining the relations. It does not consider any other information about the classes themselves. The refinement structure for RDR does not indicate any ontological refinement; however, it does indicate that an expert thought a conclusion was inappropriate and so should be replaced by another. We are looking at the possible relations between conclusions that this would allow which could the be related to the idea of relations presented here.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusion</head><p>The work we have presented is an attempt to develop techniques to discover the ontologies implicit in knowledge bases. We believe it will be of increasing importance to carry out this particular kind of knowledge discovery as larger and larger knowledge bases come into use and we seek to exploit the knowledge in the knowledge bases in different ways. We do not make any particular claim for the techniques we have developed to date, except that they suggest that such ontology discovery is possible. The key idea in the techniques we have developed is that is seems reasonable to use heuristic quantitative measures to group classes and class relations. This then enables possible ontologies to be explored on a reasonable scale</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure. 1 .</head><label>1</label><figDesc>Figure. 1. an example of a class which is a disjunction of two rules.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head></head><label></label><figDesc>Satisfactory lipid profile previous raised LDL noted &lt;--(LDL &lt;= 3.4) AND Triglyceride is NORMAL) AND (Max(LDL) &gt; 3.4) OR ((LDL is NORMAL) AND (Triglyceride is NORMAL) AND Max(LDL) is HIGH)</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 2 .</head><label>2</label><figDesc>Figure 2. Example of MCRDR tree.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 4 .</head><label>4</label><figDesc>Figure 4.Figure5Figure6</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figure 5 Figure 6 Figure 4</head><label>564</label><figDesc>Figure 4.Figure5Figure6Figure4suggests how we can find a similarity measure between 2 classes. It shows that Class X is the disjunction of nodes 1,2 and 3 and Class Y is the disjunction of nodes 4 and 5. We propose that ClassSimilarity(X,Y) = (v1 + v2 + v3) / 3, where we choose the v such that all nodes are covered by at least one edge and the sum of v (eg. v1 + v2 + v3) is maximal. Note that v stands for the Similar() function. In later similar diagrams v stands for the Subsume() and MutualEx() measures. Eg. ClassSimilarity(Go swimming, Play chess) = ((1/3)+(1/2)+(2/3))/3=0.5. We assume here the quality of all rule paths are 1.0 to simplify the example.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_7"><head></head><label></label><figDesc>For example: Subsume(A,B)==1.0, Subsume(A,C)==1.0, MutualExclusive(B,C)&gt;0.5 for all sets of three classes A,B,C from a knowledge basesWe can then combine such elements. E.g we may join Triangle(A,B,C) and Triangle(D,E,F) if A=D and MutualExclusive( {B,C}, {E,F}, &gt; 0.5). If MutualExclusive( B, C, == 1.0 ), we could say {B,C} are exhaustive subclass partitions of A<ref type="bibr" target="#b22">[23]</ref>. We applied this technique to the thyroid knowledge base and obtained the result in figure7.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_8"><head>Figure 7 .</head><label>7</label><figDesc>Figure 7. Partial taxonomy of the thyroid knowledge base</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_9"><head>[5] interpretation [ 6 ]</head><label>6</label><figDesc>interpretation [Elevated [Satisfactory prolactin level.] prolactin persists. Suggest TSH Horm one knowledge bases system sim ilarity-value &gt; 0.66 Hormone knowledge bases system subsum ption-value = 1 Horm one knowledge bases system m utual-exclusivity-value=1 taxonomy in Fig 7.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head></head><label></label><figDesc>it gives: class Go Swimming subsumes class Swimming in the beach with degree of confidence 0.83; class Play_Chess and class Go_Swimming are mutually of confidence 0.17; class Go Swimming class is similar to class Play Chess have degree of similarity 0.50. This quantitative measure enables us to group different examples of the class and provides information on whether across these examples, a class tends to subsume another class (for example, class Go Swimming subsumed class Swimming in the beach with degree of confidence 0.83. Using this quantitative measure we can say class Go Swimming tends to subsume or almost subsumes class Swimming in the beach, rather than simply saying class Go Swimming subsumes class Swimming in the beach or class Go Swimming does not subsume class Swimming in the beach.</figDesc><table><row><cell></cell><cell>Rule 1:</cell></row><row><cell></cell><cell>if true then ...</cell></row><row><cell cols="2">except</cell><cell>except</cell></row><row><cell>Rule 2:</cell><cell></cell><cell>Rule 8:</cell></row><row><cell cols="2">if sky=SUNNY then</cell><cell>if wind &gt; 30 km/h</cell></row><row><cell cols="2">Go Swimming</cell><cell>then Play chess</cell></row><row><cell cols="3">except These measures become interesting when applied to real except except</cell></row><row><cell cols="3">examples such as: class [Euthyroid levels] subsumes</cell></row><row><cell cols="3">Rule 3: class [Levels if ultraViolet=VERY HIGH then Swimming at indoor swimming pool replacement] with degree of confidence 0.866. except Rule 9: if ultraViolet=LOW then consistent with adequate Go Swimming</cell></row><row><cell cols="3">(Medically, 'adequate' thyroid hormone replacement brings</cell></row><row><cell cols="3">thyroid hormone levels to approximately normal levels.)</cell></row><row><cell>Rule 4:</cell><cell></cell><cell>Rule 10:</cell></row><row><cell cols="2">if wave = LOW then</cell><cell>if sky=RAINY then</cell></row><row><cell cols="2">Swimming in the beach</cell><cell>Play chess</cell></row><row><cell>except</cell><cell cols="2">Rule 5: if ultraViolet=MEDIUM and</cell></row><row><cell></cell><cell>wind&lt;=30 then</cell></row><row><cell></cell><cell cols="2">Swimming in the beach</cell></row><row><cell></cell><cell>except</cell></row><row><cell>Rule 6:</cell><cell></cell></row><row><cell cols="2">if wind &gt; 40 km/h</cell></row><row><cell cols="2">then Play Chess</cell></row><row><cell></cell><cell>Rule 7:</cell></row><row><cell></cell><cell cols="2">if sky=CLOUDY then</cell></row><row><cell></cell><cell>Swimming at indoor</cell></row><row><cell></cell><cell>swimming pool</cell></row></table><note>except exclusive with degree</note></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">A philosophical basis for knowledge acquisition</title>
		<author>
			<persName><forename type="first">P</forename><surname>Compton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Jansen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Knowledge Acquisition</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="241" to="257" />
			<date type="published" when="1990">1990</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title level="m" type="main">KADS A Principled Approach to Knowledge-Based System Development</title>
		<author>
			<persName><forename type="first">G</forename><surname>Schreiber</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Wielinga</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Breuker</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1993">1993</date>
			<publisher>Academic Press</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Knowledge Modeling at the Millennium (The Design and Evolution of Protégé</title>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">E</forename><surname>Grosso</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Twelfth Workshop on Knowledge Acquisition, Modeling and Management</title>
				<meeting><address><addrLine>Voyager Inn, Banff, Alberta, Canada</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1999">2000. 1999</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Knowledge Acquisition First, Modelling Later</title>
		<author>
			<persName><forename type="first">D</forename><surname>Richards</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Compton</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">10th European Knowledge Acquisition Workshop</title>
				<imprint>
			<date type="published" when="1997">1997</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">PEIRS: a pathologist maintained expert system for the interpretation of chemical pathology reports</title>
		<author>
			<persName><forename type="first">G</forename><surname>Edwards</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Pathology</title>
		<imprint>
			<biblScope unit="volume">25</biblScope>
			<biblScope unit="page" from="27" to="34" />
			<date type="published" when="1993">1993</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Ripple Donw Rules as a Mediating Representation in Interactive Induction</title>
		<author>
			<persName><forename type="first">J</forename><surname>Catlett</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Second Japanese Knowledge Acquisition for Knowledge Based Systems Workshop</title>
				<meeting>the Second Japanese Knowledge Acquisition for Knowledge Based Systems Workshop<address><addrLine>Kobe, Japan</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1992">1992</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Induction of Ripple Down Rules</title>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">R</forename><surname>Gaines</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">J</forename><surname>Compton</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Fifth Australian Conference on Artificial Intelligence</title>
				<meeting><address><addrLine>Hobart</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1992">1992</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Learning Rules with Local Exceptions</title>
		<author>
			<persName><forename type="first">J</forename><surname>Kivinen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Mannila</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Ukkonen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">European Conference on Computational Theory</title>
				<imprint>
			<date type="published" when="1993">1993</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Algebraic foundations and improved methods of induction or ripple-down rules</title>
		<author>
			<persName><forename type="first">T</forename><surname>Scheffer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">2nd Pacific Rim Knowledge Acquisition Workshop</title>
				<imprint>
			<date type="published" when="1996">1996</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Variations and Local Exception in Inductive Logic Programming</title>
		<author>
			<persName><forename type="first">A</forename><surname>Siromoney</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Siromoney</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Machine Intelligence -Applied Machine Intelligence</title>
				<editor>
			<persName><forename type="first">K</forename><surname>Furukawa</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Michie</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Muggleton</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="1993">1993</date>
			<biblScope unit="page" from="213" to="234" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">The Use of Simulated Experts in Evaluating Knowledge Acquisition</title>
		<author>
			<persName><forename type="first">P</forename><surname>Compton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Preston</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Kang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">9th AAAIsponsored Banff Knowledge Acquisition for Knowledge Base System Workshop</title>
				<meeting><address><addrLine>Canada</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1995">1995</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Simulated Expert Evaluation of Multiple Classification Ripple Down Rules</title>
		<author>
			<persName><forename type="first">B</forename><surname>Kang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Compton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Preston</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">11th Banff knowledge acqusition for knowledge-based systems workshop</title>
				<meeting><address><addrLine>Banff</addrLine></address></meeting>
		<imprint>
			<publisher>SRDG Publications</publisher>
			<date type="published" when="1998">1998</date>
		</imprint>
		<respStmt>
			<orgName>University of Calgary</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">The automatic compression of Multiple Classification Ripple Down Rule Knowledged Based Systems: Preliminary Experiments</title>
		<author>
			<persName><forename type="first">H</forename><surname>Suryanto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Richards</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Compton</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Knowledge-based Intelligence Information Engineering Systems</title>
				<meeting><address><addrLine>Adelaide, South Australia</addrLine></address></meeting>
		<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="1999">1999</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<title level="m" type="main">Transforming Rules and Trees into Comprehensible Knowledge Structures</title>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">R</forename><surname>Gaines</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1995">1995</date>
			<publisher>AAAI/MIT Press</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">From Multiple Classification RDR to Configuration RDR</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Ramadan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">11th Banff Knowledge Acquisition for Knowledge Base System Workshop</title>
				<meeting><address><addrLine>Canada</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1998">1998</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">NRDR for the Acquisition of Search Knowledge</title>
		<author>
			<persName><forename type="first">G</forename><surname>Beydoun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hoffmann</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">10th Australian Conference on Artificial Intelligence</title>
				<imprint>
			<date type="published" when="1997">1997</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">A help desk system with intelligent interface</title>
		<author>
			<persName><forename type="first">B</forename><surname>Kang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Applied Artificial Intelligence</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<biblScope unit="issue">7-8</biblScope>
			<biblScope unit="page" from="611" to="631" />
			<date type="published" when="1997">1997</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Extending Ripple Down Rules</title>
		<author>
			<persName><forename type="first">P</forename><surname>Compton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Richards</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Australian Knowledge Acquisition Workshop 99</title>
				<imprint>
			<publisher>UNSW</publisher>
			<date type="published" when="1999">1999</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<title level="m" type="main">Ounce of Knowledge is Worth a Ton of Data: Quantitative Studies of the Trade-off between Expertise and Data based on Statistically Well-founded Empirical Induction</title>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">R</forename><surname>Gaines</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1989">1989. 1989</date>
			<publisher>Morgan Kaufmann</publisher>
			<pubPlace>San Mateo, California</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Uncovering the conceptual models in RDR KBS</title>
		<author>
			<persName><forename type="first">D</forename><surname>Richards</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Compton</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Conceptual Structures ICCS&apos;97</title>
				<meeting><address><addrLine>Seattle</addrLine></address></meeting>
		<imprint>
			<publisher>Springer Verlag</publisher>
			<date type="published" when="1997">1997</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">The Reuse of Knowledge in Ripple Down Rule Knowledge Based Systems</title>
		<author>
			<persName><forename type="first">D</forename><surname>Richards</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Artificial Intelligence Department</title>
				<meeting><address><addrLine>New South Wales; Sydney</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1998">1998</date>
			<biblScope unit="page">335</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Discovery of class relations in exception structured knowledge bases</title>
		<author>
			<persName><forename type="first">H</forename><surname>Suryanto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Compton</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">The International Conference on Conceptual Structures</title>
				<imprint>
			<date type="published" when="2000">2000</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Evaluation of Taxonomic Knowledge in Ontologies and Knowledge Bases</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">G</forename><surname>Perez</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">KAW&apos;99, Twelfth Workshop on Knowledge Acquisition, Modeling andManagement. 1999</title>
				<meeting><address><addrLine>Voyager Inn, Banff, Alberta, Canada</addrLine></address></meeting>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Fast Discovery of Association Rules</title>
		<author>
			<persName><forename type="first">R</forename><surname>Agrawal</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Knowledge Discovery and Data Mining</title>
				<editor>
			<persName><forename type="first">U</forename><forename type="middle">M</forename><surname>Fayyad</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Piatetsky-Shapiro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Smyth</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="1996">1996</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
