=Paper=
{{Paper
|id=Vol-1428/BDM2I_2015_paper_8
|storemode=property
|title=Multi-Strategy Semantic Web Reasoning for Medical Knowledge Bases
|pdfUrl=https://ceur-ws.org/Vol-1428/BDM2I_2015_paper_8.pdf
|volume=Vol-1428
|dblpUrl=https://dblp.org/rec/conf/semweb/WoenselMAA15
}}
==Multi-Strategy Semantic Web Reasoning for Medical Knowledge Bases==
Multi-Strategy Semantic Web Reasoning for Medical Knowledge Bases William Van Woensel, Hossein Mohammadhassanzadeh, Samina Abidi, Raza Abidi NICHE Research Group, Faculty of Computer Science, Dalhousie University, Halifax, Canada {william.van.woensel, hassanzadeh, samina.abidi, raza.abidi}@dal.ca Abstract. Semantic Web technology offers excellent advantages for real-world medical knowledge bases, both on and off the Web. Based on ontologies from (bio)medical domains, OWL inferencing enhances knowledge bases with new facts, while deductive rules, written in semantic rule languages, supply additional inferences based on deterministic knowledge from e.g., Clinical Practice Guidelines (CPG). We argue that other mechanisms, representing weaker forms of inferencing, are also useful in dealing with incomplete healthcare knowledge. This includes inductive generalization, which leverages data similarities to induce new rules, and analogical reasoning, which relies on plausible domain knowledge. To cope with their shortcomings, we propose integrating such weak inferencing with a single, explanation-based generalization, allowing us to leverage their complementary strengths as well as apply a tutor-based paradigm for verification. In this integrated approach, justifications are generated explaining the potential correctness of queries, where missing medical knowledge is compensated by injecting plausible inferences. Based on their expertise, healthcare experts may then confirm particular justifications, materializing them in the knowledge base. Inversely, we argue that by leveraging OWL ontological knowledge, weak inferencing methods benefit from Semantic Web technology as well. Keywords: medical knowledge bases, Semantic Web reasoning, explanation- based generalization, inductive generalization, analogical reasoning 1 Introduction Clinical Decision Support Systems (CDSS) are knowledge-based systems, designed to assist health professionals in making ‘evidence-informed’ clinical decisions, based on the best evidence and approved clinical protocols [1] in concert with a patient’s medical records [2]. Given a clinical case, a CDSS reasons over the available knowledge to generate evidence-based recommendations for diagnosis/therapy, alerts, risk assessments and order sets. For the knowledge base of a CDSS, the source of knowledge are published, evidence-based clinical guidelines and the expert’s clinical experiences [3]. Developing a complete and consistent clinical knowledge base therefore is a major undertaking, as it requires (a) the computerization of the clinical guidelines in terms of decision rules, (b) the validation of the abstracted rules to ensure relevance, consistency and conciseness, and (c) implementing the validated rules in terms of an executable knowledge representation formalism. To realize medical knowledge bases, the Semantic Web framework offers a formal, logic-based framework for modeling the knowledge from clinical guidelines—i.e., via expressive biomedical OWL ontologies, allowing clinical facts to be inferred based on utilized ontological constructs; and using deductive rule languages, such as RuleML, SPIN or SWRL, which allow applying deterministic knowledge to infer clinical conclusions. From a knowledge engineering perspective, we posit that semantic ontologies are particularly useful for representing and reasoning over medical knowledge bases. This observation is reflected by work in CDSS likewise applying Semantic Web technology [4–6]. Additionally, we argue that useful mechanisms for enriching semantic medical knowledge bases are not limited to a priori, deterministic knowledge. To further extend the knowledge closure, other mechanisms can be considered as well, for instance based on data similarities and plausible knowledge (i.e., knowledge that will likely not hold in all possible situations). Inductive generalization hypothesizes that commonalities between data entities likely account for a particular shared feature; this knowledge can then be formalized in terms of rules [7]. Analogical reasoning is guided by plausible domain knowledge, implying if two data entities are similar in one specific aspect, they are likely similar in another as well [8]. These two approaches are complementary: induction is a data-intensive approach, requiring a large number of positive examples; while analogy relies on less data, but requires the presence of (plausible) knowledge. In any case, these kinds of reasoning (also including e.g., abductive reasoning [9]) are considered weaker forms of inferencing, as they do not guarantee the truth of inferred knowledge [7]. This is especially problematic in medical settings, where drawing correct, evidence-based conclusions from the available data is paramount. An effective way to cope with this tentative nature is by integrating multiple weak methods with deterministic, explanation-based generalization, allowing their complementary strengths to be leveraged [7, 10, 11]. Based on facts and deductive rules, this generalization constructs justifications that explain why a given statement may be correct [11]. In an integrated approach, weak inferencing is applied when full deterministic knowledge is lacking, yielding plausible justifications that are supplemented with weak inferences [10]. Importantly, such integrated justifications allow medical domain experts to carefully weigh the tentative data and mechanisms used in the process. The expert may then choose to confirm and fine-tune a particular justification, to be materialized in the knowledge base for future re-use. We argue that a multi-strategy reasoning approach is useful for enhancing access to medical semantic knowledge bases. In particular, we present a Semantic Web medical expert system that generates & presents plausible justifications to medical experts, using deterministic and plausible inferencing methods. This expert can further confirm and fine-tune particular justifications to add them to the knowledge base. In contrast to fully automated mechanisms, our approach combines machine inferencing capabilities with the tacit, medical experience of human experts, resulting in fully human-verified knowledge bases. Inversely, we argue that the accuracy of weak inferencing methods can be improved by leveraging Semantic Web technology, thus enhancing their utility in the knowledge engineering cycle. In particular, rich ontological knowledge can be leveraged to improve data similarity checking, as well as increase the expressivity of plausible knowledge. Our semantic expert system accepts RDF, OWL ontologies and deductive semantic rules (SPIN rules) as input, and generates plausible justifications for presentation, verification and fine-tuning by the domain expert. Confirmed justifications are transformed into RDF triples and added to the knowledge base. In Section 2, we elaborate on the reasoning layer of our expert system, discussing explanation-based generalization, inductive generalization and analogical reasoning. Section 3 details the UI providing query access to the domain expert. In Section 4, we summarize our expert system implementation. Section 5 presents related work, and Section 6 finally ends with conclusions and future work. 2 Reasoning Layer A logical first step to supplementing incomplete Semantic Web knowledge bases is to apply ontology inferencing; i.e., leveraging domain-specific, ontological knowledge to infer new information. As a second step, our integrated approach taps an additional source of deterministic knowledge, namely deductive Semantic Web rules. At its core, our approach applies explanation-based generalization, which constructs justifications explaining why a given query instance is correct. This generalization is implemented via backward chaining, which recursively finds rules proving the query statement. In this process, premises of found rules may be proven by knowledge base facts or again by other deductive rules; either ending when all involved premises have been proven, or when some premises were found to be unsatisfiable. This is illustrated in Figure 2, which visualizes the initial query justification as a tree structure. We note that for clarity, figures and rules use predicate logic notation (namespaces are omitted), and type restrictions on arguments are abbreviated as nested “(type X)” statements. In the tree, the root represents the query instance; leafs stand for facts in the knowledge base (solid line) or missing premises (dashed line); and intermediate nodes represent conjunctive rule premises, connected via directed edges to the rule conclusion. Grey-shaded leafs are facts inferred via ontology reasoning. Fig. 1. Initial justification tree. To explain our multi-strategy reasoning approach, we present a medical use case, where the system is asked to explain whether patient p01 should be recommended Azithromycin, an antibiotic typically used to treat bacterial infections. Below, we show the deductive rules used to build the initial justification tree: recommend(P, Azithromycin) : − type(P, Adult), type(P, NoLiverDiseasePatient), illness(P, ChronicBronchitis), illness(P, (type Pneumonia)) . illness(P, ChronicBronchitis) : − symptom(P, (type Cough)), symptom(P, Fatigue), symptom(P, (type Chills)), symptom(P, SlightFever) . Fig. 2. Deductive rules used in explanation-based generalization. We note that the lack of liver disease history was inferred via an OWL property cardinality restriction: cds:NoLiverDiseasePatient rdfs:subClassOf [ a owl:Restriction ; owl:maxQualifiedCardinality “0” ; owl:onProperty cds:illness ; owl:onClass cds:LiverDisease ] . Fig. 3. OWL constraint cds:NoLiverDiseasePatient. In the knowledge base, each current and past patient illness is recorded using the illness property. In case patients are not linked via the illness property to any instance of LiverDisease, this constraint implies that they are member of the cds:noLiverDiseasePatient class extension. Currently, the cough symptom and pneumonia diagnosis are missing for p01, leading to an incomplete justification. To prove the correctness of the missing premises, inductive generalization and analogy-based reasoning are applied, resulting in a plausible justification tree that incorporates multi-strategy inferencing results. We elaborate on these two methods in the subsections below. 2.1 Analogy-based reasoning Analogical inferencing supplements an incomplete knowledge base by implying that two entities who are similar in a particular aspect, are likely similar in another specific aspect as well. As such, knowledge can be transferred from a well-known entity S to a similar, lesser known entity T [8, 12, 13]. This kind of reasoning is guided by plausible rules of the form [7]: 𝑄(𝑋, 𝑌): ~𝑃(𝑋, 𝑍) Fig. 4. General structure of a plausible rule. This rule states that “Q is plausibly determined by P”, meaning if two entities S and T are both characterized by the same feature P, then they will likely also share feature Q. Below, we show a relevant plausible rule from our medical use case: illness(P, (type Pneumonia)) :~ type(P, ReducedImmunityPatient), symptom(P, (type ShortnessOfBreath)), symptom(P, (type Fever)), symptom(P, ShakingChills) Fig. 5. Example plausible rule. This rule indicates that two data entities sharing particular characteristics and symptoms, namely a reduced immune system, a type of shortness of breath, a type of fever and shaking chills, will likely share the same kind of pneumonia. This example shows how ontological knowledge may be leveraged to increase the expressivity of plausible rules. In particular, by allowing to specify rules at arbitrary levels of granularity: any kind of fever, shortness of breath or reduced immune system suffices as knowledge-transfer condition, while the patient should exhibit “shaking chills” in particular. We note that the rule only specifies which kinds of features will collectively account for a type of pneumonia; the particulars are left up to the well-known entity. Below, we show RDF statements for patient p07 and p01: cds:p07 rdf:type cds:OrganTransplantPatient ; cds:symptom cds:ShortBreathOnExertion cds:symptom cds:MediumFever ; cds:symptom cds:ShakingChills ; cds:illness cds:ViralPneumonia ; … cds:p01 rdf:type cds:HIVP\atient ; cds:symptom cds:ShortBreathOnExertion cds:symptom cds:MediumFever ; cds:symptom cds:ShakingChills ; … Fig. 6. RDF code describing entities p01 and p07. By sharing the same kind of shortness of breath (i.e., ShortBreathOnExertion), fever (i.e., MediumFever) and “shaking chills”, p01 shares almost all the required characteristics with p07. By further leveraging ontological knowledge, the system can cope with inexact matches as well; in particular, the system is able to infer that HIVPatient and OrganTransplantPatient are both subtypes of ReducedImmunityPatient. Based on the ontology hierarchy, our system can calculate and present accurate concept similarities to the medical expert. In Section 2.2, we elaborate on establishing such semantic similarities. Regarding our example, since patient p01 adheres to the knowledge-transfer condition (in italic), it can be plausibly inferred that p01 has viral pneumonia (in bold). Table 1 shows the pseudocode for analogical reasoning in our expert system. The analogical reasoning process is driven by failed justification premises. In this case, the process starts by looking for plausible rules that may resolve a missing premise (line 1). The knowledge base is then searched for facts that unify the rule; i.e., match the rule condition and consequent (line 2). For each found fact, the algorithm checks whether its instantiated knowledge-transfer condition matches the original entity (line 3). If so, its consequent is added to the entity, at least in the system’s working memory (line 4). If the expert confirms the overall justification, the working memory will be materialized in the knowledge base. 1. given failed premise, query KB for plausible rules that infer the missing property 𝒆𝒙𝒂𝒎𝒑𝒍𝒆 𝒒𝒖𝒆𝒓𝒚: 𝑖𝑙𝑙𝑛𝑒𝑠𝑠(𝑝01, (𝑡𝑦𝑝𝑒 𝑃𝑛𝑒𝑢𝑚𝑜𝑛𝑖𝑎)) → 𝑓𝑖𝑛𝑑_𝑝𝑙𝑎𝑢𝑠𝑖𝑏𝑙𝑒_𝑟𝑢𝑙𝑒𝑠(𝑖𝑙𝑙𝑛𝑒𝑠𝑠(𝑋, (𝑡𝑦𝑝𝑒 𝑃𝑛𝑒𝑢𝑚𝑜𝑛𝑖𝑎)), 𝐾𝐵) = { 𝑟𝑢𝑙𝑒𝑖 } 2. for each plausible rule, find KB facts that unify the rule 𝑢𝑛𝑖𝑓𝑖𝑐𝑎𝑡𝑖𝑜𝑛𝑠 = { ∀ 𝑟𝑢𝑙𝑒𝑖 : 𝑢𝑛𝑖𝑓𝑖 = 𝑢𝑛𝑖𝑓𝑦(𝑟𝑢𝑙𝑒𝑖 , 𝐾𝐵) } 3. for each unification, check whether its condition matches the input entity 𝑟𝑒𝑠𝑢𝑙𝑡𝑠 = { ∀ 𝑢𝑛𝑖𝑓𝑖 : 𝑟𝑒𝑠𝑢𝑙𝑡𝑖 = 𝑚𝑎𝑡𝑐ℎ𝑒𝑠(𝑐𝑜𝑛𝑑𝑖𝑡𝑖𝑜𝑛(𝑢𝑛𝑖𝑓𝑖 ), 𝑝01) } 4. add the consequent of a positive match to the input entity 𝑝01 ← 𝑐𝑜𝑛𝑠𝑒𝑞𝑢𝑒𝑛𝑡(𝑢𝑛𝑖𝑓𝑖 ) Table 1. Analogical reasoning algorithm Figure 8 illustrates part of our justification tree extended by analogical reasoning. Fig. 7. Part of justification tree extended by analogical reasoning. 2.2 Inductive generalization Inductive generalization generates new deterministic knowledge, based on similarities between facts in the knowledge base. Given a particular feature, the knowledge base is searched for a set of entities sharing the feature (called “positive coverage”). Other similarities between these entities are hereby hypothesized as accounting for the feature [7]. Applying inductive reasoning gives rise to the following type of rule: 𝑓𝑒𝑎𝑡𝑢𝑟𝑒(𝑋, 𝑣𝑎𝑙𝑢𝑒): −𝑠𝑖𝑚𝑖𝑙𝑎𝑟𝑖𝑡𝑦𝑆𝑖 (𝑋) Fig. 8. General structure of an induced rule Where feature(X, value) represents the input feature, and similaritySi represents similarities between the found entities Si. By applying this rule, the feature value is thus inferred for any entity X, which exhibits the characteristics assumed to account for the feature. We note that in this process, the number of matching entities (positive coverage) strengthens the validity of the hypothesis. Below, we discuss different ways of determining entity similarity. A good basis for finding commonalities is to take the conjunction between entity properties, i.e., all shared properties [7]. As mentioned before (see Section 3.1), conceptual similarity between entity types can be considered as well. Calculating semantic concept similarity has been studied extensively in the literature (e.g., [14, 15]). Typically, these works determine concept similarity by calculating their conceptual distance, indicated by the distance to their closest common subsuming concept. It is further argued that two “specific” concepts, lower in the concept hierarchy, will be more similar than two “abstract” concepts [15]. We rely on the well-known measure suggested by Wu and Palmer [14], which considers both conceptual distance and concept specificity: 2 × 𝑁3 𝐶𝑜𝑛𝑆𝑖𝑚(𝐶1 , 𝐶2 ) = 𝑁1 + 𝑁2 + 2 × 𝑁3 Fig. 9. Conceptual similarity calculation Where C3 is the closest common parent of C1 and C2; N1 is the number of nodes on the path from C1 to C3, and N2 is the number of nodes between C2 and C3 (indicating conceptual similarity); and N3 is the number of nodes on the path from C3 to the root (indicating conceptual specificity). In our approach, the medical expert will use this information to confirm, fine-tune or reject an inference (see Section 4). The following RDF code shows two similar entities p11 and p15, which share a number of characteristics as well as a particular feature currently missing from our justification tree (see Figure 2): cds:p11 cds:symptom cds:SlightFever ; cds:symptom cds:ChestDiscomfort cds:symptom cds:MucusProduction ; cds:symptom cds:CoughWithMucus ; … cds:p15 cds:symptom cds:SlightFever ; cds:symptom cds:ChestDiscomfort cds:symptom cds:MucusProduction ; cds:symptom cds:CoughWithMucus ; … Fig. 10. RDF code describing similar entities p11 and p15. In this case, the following deductive rule could be inferred (positive coverage of 2): symptom(P, CoughWithMucus) : − symptom(P, SlightFever), symptom(P, ChestDiscomfort), symptom(P, MucusProduction) Fig. 11. Rule generalized via inductive reasoning. We note that, depending on the facts in the knowledge base, different sets of shared characteristics may yield the same missing feature, meaning multiple deductive rules could be inferred. Table 2 shows the pseudocode for inductive reasoning in our system. As before (see Table 1), the process is triggered by a failed premise in the justification tree. Firstly, the knowledge base is queried with (a generalized version of) the failed premise, to retrieve all entities with the missing property value (line 1). Subsequently, similarity is determined between matching entities and the query entity (line 2) using the similarity function (see Figure 9). The next step is to aggregate equivalent similarities (i.e., involving the same property values), thus collecting the necessary information (e.g., positive coverage) for each potential deductive rule (line 3). Afterwards, deductive rules are generated based on these aggregated similarities (line 4). These will be presented to the domain expert (see Section 4), who can possibly choose one of the rules to be added to the knowledge base. 1. given failed premise, query KB for matching facts 𝒆𝒙𝒂𝒎𝒑𝒍𝒆 𝒒𝒖𝒆𝒓𝒚: 𝑠𝑦𝑚𝑝𝑡𝑜𝑚(𝑝01, (𝑡𝑦𝑝𝑒 𝐶𝑜𝑢𝑔ℎ)) → 𝑓𝑖𝑛𝑑_𝑚𝑎𝑡𝑐ℎ𝑖𝑛𝑔_𝑓𝑎𝑐𝑡𝑠(𝑠𝑦𝑚𝑝𝑡𝑜𝑚(𝑋, (𝑡𝑦𝑝𝑒 𝐶𝑜𝑢𝑔ℎ)), 𝐾𝐵) = { 𝑟𝑒𝑠𝑢𝑙𝑡𝑖 } 2. using similarity function, determine similarity between results and initial entity 𝑠𝑖𝑚𝑖𝑙𝑎𝑟𝑖𝑡𝑖𝑒𝑠 = { ∀ 𝑟𝑒𝑠𝑢𝑙𝑡𝑖 : 𝑠𝑖𝑚𝑖 = 𝑠𝑖𝑚𝑖𝑙𝑎𝑟𝑖𝑡𝑦(𝑟𝑒𝑠𝑢𝑙𝑡𝑖 , 𝑝01) } 3. aggregate equivalent similarities 𝑎𝑔𝑔𝑟𝑒𝑔𝑎𝑡𝑒𝑑 = { ∀ 𝑠𝑖𝑚𝑖 , 𝑠𝑖𝑚𝑗 (𝑖 ≠ 𝑗): 𝑠𝑖𝑚𝑖 = 𝑠𝑖𝑚𝑗 → 𝑎𝑔𝑔𝑟_𝑠𝑖𝑚( 𝑠𝑖𝑚𝑖 , 𝑠𝑖𝑚𝑗 ) } 4. generate deductive rules based on aggregated similarities 𝑟𝑢𝑙𝑒𝑠 = { ∀ 𝑎𝑔𝑔𝑟_𝑠𝑖𝑚𝑖 : 𝑓𝑒𝑎𝑡𝑢𝑟𝑒(𝑋, 𝑣𝑎𝑙𝑢𝑒): −𝑎𝑔𝑔𝑟_𝑠𝑖𝑚𝑖 (𝑋) } Table 2. Inductive generalization algorithm In Figure 12, we show part of the justification tree extended by induction. Fig. 12. Part of justification tree extended by inductive generalization. 3 User Interface Layer In our Semantic Web expert system, healthcare experts query the knowledge base. In Section 2, we elaborated on how multiple reasoning mechanisms cooperate in creating plausible query justifications. In this section, we illustrate how our system presents these justifications, and their plausible extensions, to the medical expert. Fig. 13. User Interface: Main query interface Figure 13 shows the main query interface, which presents the user with one or more justifications to explain a posed query, visualized as trees (see Figure 2). As shown, failed premises are highlighted in red. Upon selecting a failed premise, a list of automatic resolution options are shown (see Figure 14, right-panel), as well as a short description. After selecting the analogy resolution, the screen in Figure 15 is shown. In this screen, the top part describes the plausible rule being applied, using color coding to indicate shared variables/instances between the rule conjuncts. The bottom shows the found unification of the rule (see Section 3.1), as well as the inference that could be made by analogy. As mentioned, ontological knowledge may be leveraged to cope with inexact matches. In this case, HIVPatient and OrganTransplantPatient do not match exactly, but were found to have a direct common subsumer, namely ReducedImmunityPatient. This semantic similarity is presented to the domain expert, together with the calculated score (see Figure 10). Fig. 14. User Interface: Analogical Reasoning screen Fig. 15. User Interface: Inductive Generalization screen Figure 16 shows the inductive generalization screen. At the top, the screen indicates properties found by the system that can account for the missing premise (see Section 3.2). Below, the deductive rule is shown that may be induced, accompanied by the missing premise that the deductive rule would infer. Before confirming the inference, the domain expert can fine-tune the rule to cope with inductive bias. Firstly, the expert can select any instance or type to select any of their (super or sub) types in the related concept hierarchy, making the instance or type more general or specific. In this case, the domain expert chooses to replace the SlightFever condition by (type Fever), as any kind of fever may indicate a cough. After clicking SlightFever, a popup shows the Fever concept hierarchy, which allows the expert to choose any related concept. In addition, the expert may add (see “+” icon) or remove any premise from the rule condition. For instance, the domain expert could decide that having chest discomfort is unrelated to cough, and thus remove the premise from the condition. After accepting an automatic resolution option, the visualized justification tree (see Figure 14) is extended with the plausible inference nodes (highlighted in orange; not shown). In addition, its inferences will be added to the system’s working memory. Afterwards, once the domain expert accepts the extended plausible justification, the inferences will be materialized in the knowledge base. 4 Implementation Details Our system currently relies on Prolog to realize deductive reasoning, and to perform the unification steps during analogical reasoning. Our system further employs Aleph 1 (A Learning Engine for Proposing Hypotheses), an Inductive Logic Programming (ILP) system, to realize analogical reasoning. Custom Java code implements the conceptual similarity check discussed in Section 3.2. Regarding the expert system’s Semantic Web layer, we utilize the well-known Apache Jena framework to apply OWL reasoning on incoming RDF data. Conversion from RDF triples to Prolog and back is performed by custom Java code. Similarly, we wrote Java code to convert SPIN rules to Prolog, utilizing the Topbraid SPIN API. The user interface was implemented using the javax.swing packages, and utilizes the Prefuse graph library for visualizing the justification trees and concept hierarchies. 5 Related work To deal with incomplete (and partially incorrect) knowledge bases, Tecuci et al. [7, 10] employs justification trees to integrate inferences from multiple learning mechanisms. A three-step, tutor-based approach is suggested, ending with a sufficiently complete knowledge base ready for querying. Instances of justification trees, based on positive training examples and verified by a human tutor, are generalized and added to the knowledge base; allowing future similar inferences to be deductively entailed. Next, other instances are found and verified by the tutor, whereby positive and negative instances are used to fine-tune generalized justifications. Finally, inconsistencies (i.e., negative instances) are resolved by eliciting new knowledge (e.g., concepts). In contrast 1 http://www.cs.ox.ac.uk/activities/machlearn/Aleph/aleph.html#SEC1 to Tecuci et al., our goal is not to exhaustively extend a knowledge base until it is ready for querying. Instead, an important aim is to supply plausible justifications for individual queries on a case-by-case basis. For instance, in our medical use case, domain experts (e.g., physicians) need to be acutely aware of the rationale behind every diagnosis or drug recommendation, including any related deterministic or plausible inference. In addition, we provide the domain expert with the choice to materialize and fine-tuned plausible justifications in the knowledge base; allowing these inferences to be deductively entailed in the future. In our work, we illustrated how semantic technology can improve this kind of approach. We note that, regarding OWL reasoning, Horridge et al. [16] discuss efficiently obtaining justifications for OWL entailments. Fuzzy, probabilistic or possibilistic learning methods for dealing with uncertainty could be applied to further supplement justification trees (or initial knowledge bases). A number of works have studied specification and reasoning under uncertainty in OWL [17, 18]. However, as opposed to such learning methods, we note that inferences from inductive and analogical reasoning can be straightforwardly presented to domain experts. Studying the integration of such learning methods in our expert system, while still supplying useful explanations to domain experts, is considered future work. 6 Conclusions & Future work In this paper, we illustrated a multi-strategy reasoning approach for enhancing query access to medical knowledge bases. As opposed to traditional Semantic Web reasoning, based on a priori deterministic knowledge, this approach additionally leverages data similarities as well as plausible knowledge. In particular, our Semantic Web expert system supplies plausible query explanations to medical experts, by complementing deterministic reasoning with plausible inferences from analogical and inductive reasoning. We illustrated the usefulness of our system by a real-world, medical use case, which, in light of incomplete knowledge, involved applying both deterministic and plausible knowledge. Based on their real-world expertise, domain experts can confirm plausible justifications, and fine-tune the resulting deterministic knowledge. Finally, we showed that weak inferencing can benefit from Semantic Web technology as well, by improving similarity checking and expressivity of plausible knowledge. Future work consists of applying additional learning methods, and studying how their uncertainty results may be reflected in plausible justifications in a way that is straightforward for domain experts to understand. In the same vein, we aim to associate logical proofs with generated inferences; for instance, reflecting the inferencing steps responsible for inferred knowledge. Based on these “traces”, domain experts could choose to manually revert unsuitable materializations for certain entities. Additionally, we plan to run experiments with a large medical dataset, to study the impact of our integrated, multi-strategy reasoning approach on performance. References 1. Sim, I., Gorman, P., Greenes, R.A., Haynes, R.B., Kaplan, B., Lehmann, H., Tang, P.C.: Clinical Decision Support Systems for the Practice of Evidence-based Medicine. J. Am. Med. Inform. Assoc. 8, 527–534 (2001). 2. Raymond, L., Paré, G., de Guinea, A.O., Poba-Nzaou, P., Trudel, M.-C., Marsan, J., Micheneau, T.: Improving performance in medical practices through the extended use of electronic medical record systems: a survey of Canadian family physicians. {BMC} Med. Inf. {&} Decis. Mak. 15, 27 (2015). 3. Bates, D.W., Kuperman, G.J., Wang, S., Gandhi, T., Kittler, A., Volk, L., Spurr, C., Khorasani, R., Tanasijevic, M., Middleton, B.: Ten Commandments for Effective Clinical Decision Support: Making the Practice of Evidence-based Medicine a Reality. J. Am. Med. Inform. Assoc. 10, 523–530 (2003). 4. Iqbal, A.M., Shepherd, M.A., Abidi, S.S.R.: An Ontology-Based Electronic Medical Record for Chronic Disease Management. 44th Hawaii International International Conference on Systems Science (HICSS-44 2011), Proceedings, 4-7 January 2011, Koloa, Kauai, HI, USA. pp. 1–10 (2011). 5. Bouamrane, M.-M., Rector, A., Hurrell, M.: Using OWL Ontologies for Adaptive Patient Information Modelling and Preoperative Clinical Decision Support. Knowl. Inf. Syst. 29, 405–418 (2011). 6. Rodriguez-Gonza¡lez, A., Labra-Gayo, Jose Emilio Colomo-Palacios, R., Mayer, M., Gomez-Berbas, J.M., Garcia-Crespo, A.: SeDeLo: Using Semantics and Description Logics to Support Aided Clinical Diagnosis. J. Med. Syst. 36, 2471–2481 (2012). 7. Tecuci, G.: Building Knowledge Bases through Multistrategy Learning and Knowledge Acquisition. Mach. Learn. Knowl. Acquis. Integr. Approaches. (1995). 8. Gentner, D.: The Mechanisms of Analogical Reasoning. Readings Mach. Learn. (1990). 9. Henson, C.A., Thirunarayan, K., Sheth, A.P.: An Efficient Bit Vector Approach to Semantics-Based Machine Perception in Resource-Constrained Devices. 11th International Semantic Web Conference. pp. 149–164. Springer (2012). 10. Tecuci, G.: Plausible Justification Trees: A Framework for Deep and Dynamic Integration of Learning Strategies. Mach. Learn. 11, 237–261 (1993). 11. Danyluk, A.P.: The use of explanations for similarity-based learning. IJCAI. 274–276 (1987). 12. Carbonell, J.G.: Derivational Analogy: a Theory of Reconstructive Problem Solving and Expertise Acquisition. Mach. Learn. An Artif. Intell. approach. 2, (1986). 13. Winston, P.H.: Learning and Reasoning by Analogy. Commun. ACM. 23, 689–703 (1980). 14. Wu, Z., Palmer, M.: Verbs semantics and lexical selection. Proceedings of the 32nd annual meeting on Association for Computational Linguistics. pp. 133–138. Association for Computational Linguistics (1994). 15. Chiabrando, E., Likavec, S., Lombardi, I., Picardi, C., Dupré, D.T.: Semantic similarity in heterogeneous ontologies. In: Bra, P. De and Grønbæk, K. (eds.) HT’11, Proceedings of the 22nd {ACM} Conference on Hypertext and Hypermedia, Eindhoven, The Netherlands, June 6-9, 2011. pp. 153–160. ACM (2011). 16. Horridge, M., Parsia, B., Sattler, U.: Extracting Justifications from Bioportal Ontologies. Proceedings of the 11th International Conference on The Semantic Web - Volume Part II. pp. 287–299. Springer-Verlag, Berlin, Heidelberg (2012). 17. Stoilos, G., Simou, N.: Uncertainty and the semantic web. Intell. Syst. …. (2006). 18. Ding, Z., Peng, Y.: A Probabilistic Extension to Ontology Language OWL. Proceedings of the 37th Annual Hawaii International Conference on System Sciences (HICSS’04). p. 40111.1–. IEEE Computer Society, Washington, DC, USA (2004).