<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Towards the Evaluation of the LarKC Reasoner Plug-ins</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Zhisheng</forename><surname>Huang</surname></persName>
							<email>huang@cs.vu.nl</email>
							<affiliation key="aff0">
								<orgName type="department">Computer Science Department</orgName>
								<orgName type="institution">Vrije Universiteit Amsterdam</orgName>
								<address>
									<country key="NL">The Netherlands</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Towards the Evaluation of the LarKC Reasoner Plug-ins</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">81E10256389DDFEADEF1AFD2280E325D</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T00:50+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Reasoning</term>
					<term>Evaluation</term>
					<term>Benchmarking</term>
					<term>Web Scale Reasoning</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In this paper, we present an initial framework of evaluation and benchmarking of reasoners deployed within the LarKC platform, a platform for massive distributed incomplete reasoning that will remove the scalability barriers of currently existing reasoning systems for the Semantic Web. We discuss the evaluation methods, measures, benchmarks, and performance targets for the plug-ins to be developed for approximate reasoning with interleaved reasoning and selection. In this paper, we propose a specification language of gold standards for the evaluation and benchmarking, and discuss how it can be used for the evaluation of reasoner plug-ins within the LarKC platform.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>The essence of the LarKC project <ref type="bibr" target="#b1">[2]</ref> <ref type="foot" target="#foot_0">1</ref> is to go beyond notions of absolute correctness and completeness in reasoning. We are looking for retrieval methods that provide useful responses at a feasible cost of information acquisition and processing. Therefore, generic inference methods need to be extended to nonstandard approaches. In consequence, traditional metrics such as completeness and correctness for reasoning need to be replaced by metrics that ratio the utility of solutions with their related costs as a means to evaluate the chosen problem solver. Therefore, we will develop a framework for evaluation and measuring the relative utility of various reasoning approaches that will be implemented in the LarKC project.</p><p>Approximate reasoning is non-standard reasoning which is based on the idea of sacrificing soundness or completeness for a significant speed-up in reasoning. This is to be done in such a way that the loss of correctness is at least outweighed by the obtained speed-up <ref type="bibr" target="#b7">[8]</ref>. Anytime reasoning in which more answers can be obtained over time is expected to be a behavior of approximate reasoning for the LarKC platform. Interleaving reasoning and selection is considered to be an approach to improving the performance of the LarKC platform <ref type="bibr" target="#b6">[7,</ref><ref type="bibr" target="#b3">4]</ref>. The main idea of the interleaving framework is to use selectors to select only limited and relevant part of data for reasoning, so that the efficiency and the scalability of reasoning can be improved. Those non-standard reasoning approaches need new metrics and frameworks for the evaluation and benchmarking of reasoners developed within the LarKC platform.</p><p>These new reasoning paradigms, which fuse approaches from many different fields will also require new approaches to evaluation. Traditional measures like soundness and completeness will have to be enriched with measures such as recall and precision, and worst-case complexity will have to be enriched by approaches such as anytime performance profiles. In this paper, we will develop such new evaluation measures and apply them to the implemented resoner plug-ins, both on synthetic datasets and on datasets from the use-cases.</p><p>In this paper, we will present an initial framework of evaluation and benchmarking of reasoners deployed with the LarKC platform. We will discuss the evaluation methods, measures, benchmarks, and performance targets for the plug-ins to be developed for the task of approximate reasoning with interleaved reasoning and selection. Furthermore, we will develop a specification language of gold standards for the evaluation and benchmarking.</p><p>The rest of the paper is organized as follows. In Section 2, we overview the LarKC Platform and the general picture of reasoner plug-ins, which will be developed or deployed within the LarKC platform. In Section 3, we develop a framework of evaluation and benchmarking of reasoners. In Section 4, we propose a specification language of gold standards for the evaluation and benchmarking within the LarKC platform. Section 5 discusses the related work before concluding the paper in Section 6.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">LarKC Platform</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">LarKC Architecture</head><p>In <ref type="bibr" target="#b8">[9]</ref>, the first version of the LarKC architecture has been proposed. This design is based on a thorough analysis of the requirements and considering the lessons learned during the first year of the project. Figure <ref type="figure" target="#fig_0">1</ref> shows a detailed view of the LarKC Platform architecture.</p><p>The LarKC platform has been designed in a way so that it is as lightweight as possible, but provides all necessary features to support both users and plugins. For this purpose, the following components are distinguished as part of the LarKC platform:</p><p>-Plug-in API: defines interfaces for plug-ins and therefore provides support for interoperability between platform and plug-ins and between plug-ins. -Data Layer API: provides support for data access and management.</p><p>-Plug-in Registry: contains all necessary features for plug-in registration and discovery -Workflow Support System: provides support for plug-in instantiation, through the deployment of plug-in managers, and for monitoring and controlling plug-in execution at workflow level. -Queues: provides support for deployment and management of the communication between platform and plug-ins and between plug-ins.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">LarKC resoner plug-ins</head><p>Reasoning APIs All LarKC plug-ins share a common super class, namely the Plug-in class. This class provides an interface for functions common to all plug-in types. All plug-ins are identified by a name, which is a string. Plugins provide meta data that describes the functionality that they offer. Plug-ins provide Quality of Service (QoS) information regarding how they perform the functionality that they offer. The resoner plug-in will execute a given SPARQL Query against a Triple Set provided by a Select plug-in. The interface of the resoner plug-in can be seen in Table <ref type="table" target="#tab_0">1</ref>.</p><p>The resoner plug-in supports the four standard methods for a SPARQL endpoint, namely select, describe, construct, and ask. The input to each of the reason methods are the same and includes the query to be executed, the statement set to reason over, the contract, which defines the behavior of the reasoner, and the context, which defines the special information of the reasoning task. The output of these reasoning methods depends on the reasoning task being performed. The select method returns a Variable Binding as output where the variables correspond to those specified in the query. The construct and describe methods return RDF graphs, in the first case this graph is constructed according to the query and in the second the graph contains triples that describe the variable specified in the query. Finally ask returns a Boolean Information Set as output, which is true if the pattern in the query can be found in the Triple Set, or false if not.</p><p>Reasoner Plug-ins The LarKC reasoner plug-ins can range from the reasoners which provide the standard reasoning support with RDF/RDFS/OWL data to the reasoners which realize non-standard reasoning tasks such as reasoning with inconsistent ontologies, rule-based reasoning, stream reasoning. Here is an (incomplete) list of the LarKC resoner plug-ins which have been developed for the LarKC platform.</p><p>-SPARQL Query Evaluation Reasoner: This resoner plug-in wraps OWL-IM and enables the execution of SPARQL Select, Construct, Describe and Ask queries to be executed against it. -Pellet Reasoner: This resoner plug-in is a wrapper of Pellet SPARQL DL Reasoner<ref type="foot" target="#foot_2">2</ref> , which provides the reason support of Description Logics. -DIG Interface: This resoner plug-in provides the support for the DIG interface<ref type="foot" target="#foot_3">3</ref> , which allows an external DIG reasoner to be called, like RACER, FACT++,KAON2, PION, etc. -OWLAPI Reasoner: This resoner plug-in provides the support for OWL APIs, which is a standard for reasoners with OWL data. -PION Reasoner: This is a reasoner which can be used for reasoning with inconsistent ontologies. Namely, given an inconsistent ontology and a query, the PION reasoner can return a meaningful answer.</p><p>3 A Framework of Evaluation and Benchmarking for Ontology Reasoners</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">General Consideration</head><p>In this section, we will present a framework of evaluation and benchmarking for ontology reasoners. The main idea is to use the framework, which have developed in the KnowledgeWeb project for benchmarking inconsistency reasoners in the Semantic Web <ref type="bibr" target="#b5">[6]</ref>. In LarKC, we can use this framework to evaluate and benchmark both standard reasoner plug-ins and non-standard reasoner plug-ins, e.g. the PION reasoner <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b2">3]</ref> for reasoning with inconsistent ontologies.</p><p>In ontology engineering, evaluation and benchmarking target software products, tools, services, and processes. The objects are called tested systems. Evaluation and benchmarking are the systematic determination of merit, worth, and significance of tested systems. Those merit, worth, and significance are characterized as a value relation, which is considered as a preference relation, i.e., a partial order set A, . We consider a tested system as one which is targeted by the objectives of evaluation or benchmarking. A tested system can be characterized as an input-output function, alternatively, called a characteristic function of the tested system. Namely, it maps a tuple of the input parameters into an output value. A value relation is defined as a preference relation on a set of values. Namely, a value relation is characterized as a partial order set.</p><p>We define evaluation as the systematic determination of the values of tested systems with respect to its partial ordered value relation, whereas benchmarking as a continuous process for improving by systematically evaluating tested systems, and comparing them to those considered to be the best. Namely, benchmarking is the continuous process of evaluation.</p><p>In the LarKC project, Task 4.7.1 is requested to develop an initial framework for evaluation of reasoner plug-ins. For the purpose of continuous improvement of the LarKC platform, the issue of benchmarking reasoner plug-ins should be also covered.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Goals and Criteria for Evaluation and Benchmarking of Reasoner Plug-ins</head><p>We consider the following initial goals for evaluating LarKC Reasoner plug-ins:</p><p>-Bug Detection: A good evaluation of reasoner plug-ins should be able to detect hidden bugs in the implementation. These bugs may be hard to detect with manual examination by developers. It requires that test data sets cover many functionalities/use cases of reasoner plug-ins. -Robustness: A robust reasoner plug-in should not fail with noisy or erroneous test data. Thus, special test data sets should be designed to test the robustness of a reasoner plug-in. -Performance analysis: One of the main concerns on the quality of a reasoner plug-in is its performance. Thus, a necessary procedure of evaluation and benchmarking of reasoner plug-ins is to provide an analysis of their performance. The usual criteria for examining the performance of reasoner plug-ins are: (i) the time costs, including the time cost for getting the first query answer with anytime behavior, and the average time cost for each query answer, (ii) the resource consumption, including the maximal working memory request, and (iii) the quality of the query answers, which will be discussed in the next section. -Scalability Potential: The LarKC platform is expected to support Web scale reasoning. Thus, the scalability of a reasoner plug-in becomes a crucial issue for the performance of the overall platform. The scalabililty potential of a reasoner is how well it can deal with large amount of data. -Platform Improvement: A useful evaluation and benchmarking of reasoner plug-ins should be able to find bottle necks within the platform. It would provide an analysis of how the design of platform can be improved.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Measuring the Quality of Query Answers</head><p>As discussed in the last section, the quality of query answers is one of the main criteria for evaluating and benchmarking of reasoners.</p><p>The answer value set for standard ontology reasoning is usually considered as a Boolean value set, namely, it consists of 'true' and 'false'. The answer value set for reasoning with inconsistent ontologies usually consists of three values accepted, rejected, and undetermined, as introduced in the PION system. We will develop gold standards, which represents intuitive answers from a human for queries on reasoning with consistent or inconsistent ontologies. Thus, we can compare the answers from the tested system/approach with the gold standard, which is supposed to be intuitively true by a human to see to what quality of query answers provided by tested systems.</p><p>For a query with an inconsistent ontology, there might exist the following difference between an answer from the tested system/approach and its intuitive answer in a gold standard.</p><p>-Intended Answer: the system's answer is the same as the intuitive answer; -Counter-intuitive Answer: the system's answer is opposite to the intuitive answer. Namely, the intuitive answer is 'accepted' whereas the system's answer is 'rejected', or vice versa. -Cautious Answer: The intuitive answer is 'accepted' or 'rejected', but the system's answer is 'undetermined'. -Reckless Answer: The system's answer is 'accepted' or 'rejected' whereas the intuitive answer is 'undetermined'. We call it a reckless answer, because under this situation the system returns just one of the possible answers without seeking other possibly opposite answers, which may lead to 'undetermined'.</p><p>Therefore, a value set {intended answer, cautious answer, reckless answer, counter intuitive answer}, can be introduced for the evaluation of answers with gold standards. An intended answer is considered as a best one, whereas a counter intuitive answer is considered as a worse one. Cautious answers are usually not considered as wrong answers, whereas reckless answers may give wrong answers. Thus, a preference relation on the value set can be like this:</p><p>{intended answer cautious answer, cautious answer reckless answer, reckless answer counter intuitive answer} Based on this preference order, we can measure the quality of query answers by the following answer rates:</p><p>-IA Rate, which counts only intended answers. Namely the Intended Answer Rate is defined as the ratio of the amount of Intended Answers to the total amount of the answers. -IC Rate, which counts non-error answers. Namely, IC Rate = (Intended Answers +Cautious Answers)/TotalAnswerNumber.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4">Workflows of Evaluation and Benchmarking</head><p>Common data sets and common gold standards are usually used for an evaluation of different tested systems. Those systems may be heterogeneous with respect to their input data. For example, a reasoner may support only OWL data, whereas another reasoner may support only DIG data. Therefore a data translator is needed to convert data sets represented in a standard format into the data sets which are represented in a format that is supported by a tested system. Based on a comparison between test results and gold standards, result evaluation can be done manually, semi-automatically, or automatically. The output of the result The methods of statistics and visualization are usually introduced in this phase for better illustration. The evaluation results will be ranked with respect to its value relation. Finally, it leads to an evaluation report which concludes the values of tested systems and explain the reasons why the system behaves differently.</p><p>An investigation is usually made to detect the problem of tested systems based on the analysis of the evaluation. The workflow of evaluation is shown in Figure <ref type="figure" target="#fig_1">2</ref>.</p><p>As discussed above, benchmarking is a continuous processing of evaluation. Therefore for benchmarking, evaluation results are used further for the improvement of tested systems. This would usually lead to new versions of tested systems. Based on a benchmarking analysis, new test data sets may be re-designed or previous data sets are adjusted for further evaluation with respect to some targeted problems. The workflow of benchmarking is shown in Figure <ref type="figure" target="#fig_2">3</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">A Specification Language for Gold Standards</head><p>Manual evaluation and analysis of test results are usually time consuming, labor intensive, and error prone. The formalism of gold standard will pave a way for automatic or semi-automatic evaluation and analysis of test results.</p><p>A gold standard is an evaluation function which maps queries into answers with confidence values. For reasoner benchmarking, a gold standard is a (partial) function which maps queries into (intuitive) answers with a confidence value. For example, for benchmarking inconsistency processing, we considered the answer set {accepted, rejected, undetermined}. For a query "are birds animals?", the expected answer is intuitively considered as "accepted" with confidence value "1.0". However, for the query "are men animals?", the expected answers may be well suitable to be specified as an answer lower confidence value, say, "ac- xsi:schemaLocation="http://wasp.cs.vu.nl/larkc/d471/gd.xsd"&gt; &lt;name value="LarKC gold standard example 1" version="0.0.1"/&gt; &lt;comment text="just an example, which is independent from any particular ontology. It is up to evaluators to decide which ontology can be applied"/&gt; &lt;query id="Are birds animals?" querytype="subsumes"&gt; &lt;queryBody&gt; ... &lt;/queryBody&gt; &lt;expectedAnswers&gt; &lt;answer value="accepted" confidence="1"/&gt; &lt;/expectedAnswers&gt; &lt;/query&gt; &lt;query id="Are men animals?" querytype="subsumes"&gt; &lt;queryBody&gt; ... &lt;/queryBody&gt; &lt;expectedAnswers&gt; &lt;answer value="accepted" confidence="0.4"/&gt; &lt;answer value="undetermined" confidence="0.2"/&gt; &lt;answer value="rejected" confidence="0.4"/&gt; &lt;comment text="just an example which shows the possibility of multiple answers in a gold standard"/&gt; &lt;/expectedAnswers&gt; &lt;/query&gt; &lt;/goldenStandard&gt; cepted" with confidence value "0.4", "rejected" with confidence value "0.4", and "undetermined with the confidence value "0.2". Namely, we use the confidence values to represent some kinds of uncertainty of expected answers. The confidence values can be obtained by various approaches, like from questionnaires, statistics, machine learning, etc.</p><p>We design gold standards which are independent from a specific ontology. Namely, it is up to evaluators/users to decide which ontologies can be applied with respect to a gold standard.</p><p>In the following, we develop a gold standard specification language which is suitable for SPARQL queries as reasoning queries. Thus, it is an XML file, which is easy to use and read. Table <ref type="table" target="#tab_1">2</ref> shows an example of a gold standard which is encoded as an XML document.</p><p>This XML document specifies the name and the version of the gold standard. Each query consists of a detailed query statement (in the SPARQL query language) and its expected answer specification. Each expected answer is attached by a confidence value. For non-Boolean answers, like those for sparqlSelect and sparqlConstruct queries which would return a variable binding or a rdf graph as an answer, the values of expected answers are specified with a detailed xmlencoded subtree as specified in Table <ref type="table" target="#tab_2">3</ref>.</p><p>Alternatively, the gold standard specification language can be defined by using the standard meta data languages or ontology language, such as RDF/RDFS/OWL. Namely, a gold standard can be specified as meta data or an ontology, which provides a possibility for reasoning with gold standards. Table <ref type="table" target="#tab_3">4</ref> shows an example of the RDF representation of the gold standards.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Related Work</head><p>In <ref type="bibr" target="#b0">[1]</ref>, Castro develops an evaluation and benchmarking methodology for Semantic Web technologies and presents various methods for RDF(S) and OWL interoperatability benchmarking.</p><p>SEALS 4 is a project on Semantic Evaluation at Large Scale. The goal of the SEALS project is to provide an independent, open, scalable, extensible, and sustainable evaluation infrastructure for semantic technologies. The SEALS Platform allows the remote evaluation of semantic technologies thereby providing an objective comparison of the different existing semantic technologies. The SEALS Platform will provide an integrated set of semantic technology evaluation services and test suites. Therefore, one of the future work is to integrate the evaluation methods and benchmarks developed in the context of LarKC with the SEALS Platform.</p><p>The work on the evaluation design for advanced reasoning systems in the SEALS project <ref type="bibr" target="#b9">[10]</ref> is still under development. We have observed that the SEALS project has presented the definition of the evaluations and test data that will be used in the first SEALS Evaluation Campaign. The tests have been designed to cover the interoperability and the performance of advance reasoning systems.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Conclusions</head><p>In this paper, we have presented an initial framework of evaluation and benchmarking for reasoner plug-ins within the LarKC platform. We have proposed the evaluation methods, measures, benchmarks, and performance targets for the plug-ins to be developed for approximate reasoning with interleaved reasoning and selection. Based on the initial framework, we have discussed the workflows of evaluation and benchmarking. Furthermore, in this paper, we have proposed a specification language of gold standards for evaluation and benchmarking. We have discussed how the proposed language of gold standards can be used for the evaluation of reasoner plug-ins within the LarKC platform.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. The LarKC Platform Architecture</figDesc><graphic coords="3,137.60,115.84,340.15,249.06" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. Workflow of evaluation.</figDesc><graphic coords="7,165.94,115.84,283.47,168.75" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Fig. 3 .</head><label>3</label><figDesc>Fig. 3. Workflow of benchmarking</figDesc><graphic coords="8,165.95,115.84,283.46,158.96" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 .</head><label>1</label><figDesc>Reasoner Plug-in Interface</figDesc><table><row><cell>Function name</cell></row><row><cell>sparqlSelect(SPARQLQuery q, SetOfStatements s, Contract c, Context ctx)</cell></row><row><cell>sparqlConstruct(SPARQLQuery q, SetOfStatements s, Contract c, Context ctx)</cell></row><row><cell>sparqlDescribe(SPARQLQuery q, SetOfStatements s, Contract c, Context ctx)</cell></row><row><cell>sparqlAsk(SPARQLQuery q, SetOfStatements s, Contract c, Context ctx)</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 .</head><label>2</label><figDesc>Example of Gold Standard</figDesc><table><row><cell>&lt;?xml version="1.0" encoding="ISO-8859-1"?&gt;</cell></row><row><cell>&lt;goldenStandard xmlns="http://wasp.cs.vu.nl/larkc/d471/lang"</cell></row><row><cell>xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3 .</head><label>3</label><figDesc>XML Format for Expected Answers</figDesc><table><row><cell>...</cell></row><row><cell>&lt;query id="List all the subconcepts fo wine"</cell></row><row><cell>querytype="sparqlSelect"&gt;</cell></row><row><cell>&lt;queryBody&gt;</cell></row><row><cell>&lt;sparqlPrefix name="rdfs"</cell></row><row><cell>value="http://www.w3.org/2000/01/rdfschema#/&gt;</cell></row><row><cell>&lt;sparqlPrefix name="wine"</cell></row><row><cell>value="http://www.w3.org/TR/2003/PR-owl-guide-20031209/wine#"/&gt;</cell></row><row><cell>&lt;sparqlBody value="SELECT ?X WHERE {?X rdfs:subClassOf wine:Wine.}"/&gt;</cell></row><row><cell>&lt;/queryBody&gt;</cell></row><row><cell>&lt;expectedAnswers&gt;</cell></row><row><cell>&lt;answer confidence="1.0"/&gt;</cell></row><row><cell>&lt;value&gt;......&lt;/value&gt;</cell></row><row><cell>&lt;/expectedAnswers&gt;</cell></row><row><cell>...</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4 .</head><label>4</label><figDesc>RDF representation of the Gold standards</figDesc><table><row><cell>&lt;?xml version="1.0" encoding="UTF-8"?&gt;</cell></row><row><cell>&lt;rdf:RDF xmlns:fields="http://sindice.com/vocab/fields#"</cell></row><row><cell>......</cell></row><row><cell>xmlns:rdfs="http://www.w3.org/2000/01/rdf-schema#"&gt;</cell></row><row><cell>&lt;rdf:Description</cell></row><row><cell>rdf:about="http://wasp.cs.vu.nl/larkc/d471/lang#GoldStandardName"&gt;</cell></row><row><cell>&lt;larkc:name&gt;LarKC gold standard example 1&lt;/larkc:name&gt;</cell></row><row><cell>&lt;/rdf:Description&gt;</cell></row><row><cell>&lt;rdf:Description</cell></row><row><cell>rdf:about="http://wasp.cs.vu.nl/larkc/d471/lang#Query"&gt;</cell></row><row><cell>&lt;larkc:queryID&gt;Are birds animals?&lt;/larkc:queryID&gt;</cell></row><row><cell>&lt;larkc:queryType&gt;subsumes&lt;/larkc:queryType&gt;</cell></row><row><cell>&lt;larkc:queryBody&gt;...&lt;/larkc:queryBody&gt;</cell></row><row><cell>&lt;larkc:expectedAnswers&gt;</cell></row><row><cell>&lt;rdf:Bag&gt;&lt;rdf:li&gt;&lt;larkc:answer&gt;</cell></row><row><cell>&lt;larkc:value&gt;accepted&lt;/larkc:value&gt;</cell></row><row><cell>&lt;larkc:confidence&gt;1&lt;/larkc:confidence&gt;&lt;/larkc:answer&gt;&lt;/rdf:li&gt;</cell></row><row><cell>&lt;/rdf:Bag&gt;</cell></row><row><cell>&lt;/larkc:expectedAnswers&gt;</cell></row><row><cell>&lt;/rdf:Description&gt;</cell></row><row><cell>......</cell></row><row><cell>&lt;/rdf:RDF&gt;</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">http://www.larkc.eu Proceedings of the International Workshop on Evaluation of Semantic Technologies (IWEST</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2010" xml:id="foot_1">). Shanghai, China. November 8, 2010.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_2">http://clarkparsia.com/pellet</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_3">http://dig.sourceforge.net/</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m" type="main">Benchmarking Semantic Web Technology</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">G</forename><surname>Castro</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2010">2010</date>
			<publisher>IOS Press</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Towards larkc: A platform for web-scale reasoning</title>
		<author>
			<persName><forename type="first">D</forename><surname>Fensel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Van Harmelen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Andersson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Brennan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Cunningham</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">D</forename><surname>Valle</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Fischer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kiryakov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>School</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Tresp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Wesner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Witbrock</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Zhong</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the International Conference on Semantic Computing</title>
				<meeting>the International Conference on Semantic Computing</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="524" to="529" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Using semantic distances for reasoning with inconsistent ontolgies</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Van Harmelen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 7th International Semantic Web Conference (ISWC2008)</title>
				<meeting>the 7th International Semantic Web Conference (ISWC2008)</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m" type="main">D4.3.2 -implementation of plug-ins for interleaving reasoning and selection</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Van Harmelen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Schlobach</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Tagni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Teije</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zeng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Zhong</surname></persName>
		</author>
		<ptr target="http://www.larkc.eu/deliverables/" />
		<imprint>
			<date type="published" when="2010-03">March 2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Reasoning with inconsistent ontologies</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Van Harmelen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ten Teije</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the International Joint Conference on Artificial Intelligence -IJCAI&apos;05</title>
				<meeting>the International Joint Conference on Artificial Intelligence -IJCAI&apos;05</meeting>
		<imprint>
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Benchmarking the processing of inconsistent ontologies</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Volker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Ji</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Stuckenschmidt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Meilicke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Schlobach</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Van Harmelen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lam</surname></persName>
		</author>
		<idno>4/d2.1.6.3</idno>
		<ptr target="http://wasp.cs.vu.nl/knowledgeweb/D2163/kweb2163.pdf" />
	</analytic>
	<monogr>
		<title level="j">knowledgeweb</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="issue">2</biblScope>
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m" type="main">D4.3.1 -strategies and design for interleaving reasoning and selection of axioms</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zeng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Schlobach</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Den Teije</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Van Harmelen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Zhong</surname></persName>
		</author>
		<ptr target="http://www.larkc.eu/deliverables/" />
		<imprint>
			<date type="published" when="2009-09">September 2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">What is approximate reasoning?</title>
		<author>
			<persName><forename type="first">S</forename><surname>Rudolph</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Tserendorj</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Hitzler</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of RR2008</title>
				<meeting>RR2008</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="volume">5341</biblScope>
			<biblScope unit="page" from="150" to="164" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><surname>Witbrock</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Fortuna</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Bradesko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kerrigan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Bishop</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Van Harmelen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Teije</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Oren</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Momtchev</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Tenschert</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Cheptsov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Roller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Gallizo</surname></persName>
		</author>
		<ptr target="http://www.larkc.eu/deliverables/" />
		<title level="m">Larkc deliverable d5.3.1 -requirements analysis and report on lessons learned during prototyping</title>
				<imprint>
			<date type="published" when="2009-06">June 2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">of test data for advanced reasoning systems</title>
		<author>
			<persName><forename type="first">M</forename><surname>Yatskevich</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Stoilos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Horrocks</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Martin-Recuerda</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Seals deliverable d11.2, evaluation design and collection</title>
				<imprint>
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
