<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Evaluating Semantic Web Service Matchmaking Effectiveness Based on Graded Relevance</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Ulrich</forename><surname>Küster</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Birgitta</forename><surname>König-Ries</surname></persName>
						</author>
						<author>
							<affiliation key="aff0">
								<orgName type="department">Institute of Computer Science</orgName>
								<orgName type="institution">Friedrich-Schiller-University Jena</orgName>
								<address>
									<postCode>D-07743</postCode>
									<settlement>Jena</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff1">
								<orgName type="department">Semantic Web Service Challenge: http</orgName>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff2">
								<orgName type="department">S3 Contest on Semantic Service Selection</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">Evaluating Semantic Web Service Matchmaking Effectiveness Based on Graded Relevance</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">9828E3FBF7F53BC5C7A335899E5E4E69</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T06:33+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Semantic web services (SWS) promise to take service oriented computing to a new level by allowing to semi-automate timeconsuming programming tasks. At the core of SWS are solutions to the problem of SWS matchmaking, i.e., the problem of comparing semantic goal descriptions with semantic offer descriptions to determine services able to fulfill a given request. Approaches to this problem have so far been evaluated based on binary relevance despite the fact that virtually all SWS matchmakers support more fine-grained levels of match. In this paper, a solution to this discrepancy is presented. A graded relevance scale for SWS matchmaking is proposed as are measures to evaluate SWS matchmakers based on such graded relevance scales. The feasibility of the approach is shown by means of a preliminary evaluation of two hybrid OWL-S matchmakers based on the proposed measures.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>In recent years, semantic web services (SWS) research has emerged as an application of the ideas of the semantic web to the service oriented computing paradigm <ref type="bibr" target="#b0">[1]</ref>. The grand vision of SWS is to have a huge online library of component services available, which can be discovered and composed dynamically based upon their formal semantic annotations. One of the core problems in the area concerns SWS matchmaking, i.e. the problem of comparing a set of semantic service advertisements with a semantic request description to determine those services that are able to fulfill the given request. A variety of competing approaches to this problem has been proposed <ref type="bibr" target="#b1">[2]</ref>. However, the relative strengths and shortcomings of the different approaches are still largely unknown. For the future development of the area it is thus of crucial importance to establish sound and reliable evaluation methodologies. The recent formation of international SWS evaluation campaigns 1 is a promising step in this direction.</p><p>One of the core problems of SWS matchmaking is that it is unrealistic to expect advertisements and requests to be either a perfect match or a complete fail. Thus, virtually all SWS matchmakers support multiple degrees of match, i.e. they classify the set of advertisements into a hierarchy of different match levels or even assign a continuous degree of match to each offer. Nevertheless, existing approaches for the evaluation of the retrieval effectiveness of matchmaking approaches have so far been based exclusively on binary relevance, i.e. for evaluation purposes an advertisement is considered to be either a match or not, but no further distinction is made. This is a remarkable discrepancy that may distort evaluation results and compromise their reliability. This paper presents an approach to overcome this problem.</p><p>The rest of the paper is structured as follows. In the following Section, we provide information about related previous work. In Section 3, we discuss the notion of relevance in the domain of SWS matchmaking and propose a graded relevance scale customized to this domain. In Section 4, we introduce a number of evaluation measures capable to deal with graded relevance. In Section 5, we report on a preliminary experiment on applying the graded relevance scale and the evaluation measures to evaluate two OWL-S matchmakers. We discuss our results with a particular focus on the influence of switching measures and definitions of relevance. Finally, in Section 6, we draw conclusions and outline aspects of future work.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Related Work</head><p>Experimental evaluation of SWS retrieval has received very little attention so far. The few approaches that were thoroughly evaluated so far exclusively relied on binary relevance and standard measures based on precision and recall. This was also the case with the first edition of the S3 Contest on Semantic Service Selection <ref type="foot" target="#foot_0">2</ref> .</p><p>The first approach, and the only that we are aware of, to apply graded relevance in SWS retrieval evaluation is the work by Tsetsos et al. <ref type="bibr" target="#b2">[3]</ref>. They propose to use a relevance scale based on fuzzy linguistic variables and the application of a fuzzy generalization of recall and precision that evaluates the degree of correspondance between the rating (not ranking) of a service by an expert and a system under evaluation. In this aspect this measure is very similar to the ADM (average distance measure) measure proposed by <ref type="bibr" target="#b3">[4]</ref>. Unlike measures that evaluate the ranking created by a retrieval system these measures evaluate the absolute score assigned to a retrieved item by the system. This can lead to counterintuitive results since such measures are obviously biased against systems that rank services correctly but generally assign relatively higher or lower scores <ref type="bibr" target="#b4">[5]</ref>. The measures that we use in this work avoid this issue.</p><p>Di Noia et al. obtained reference rankings for service matchmaking evaluations by directly asking human assessors to rank the available services <ref type="bibr" target="#b5">[6]</ref>. This approach avoids the imprecision related to binary relevance judgments and generally yields more stable results than inducing a reference ranking via relevance judgments. However, it also requires much more effort from the human assessors and is thus difficult to scale to large datasets. Di Noia et al. evaluate the matchmaking performance using rank correlation measures from statistics. These measures estimate the difference between two rankings but, for instance, do not differentiate whether the rankings differ in the top ranks or the bottom ranks. Yet, for most retrieval settings, the correctness of the top ranks is much more important than that of the bottom ranks. The measures proposed in this work allow to take such considerations into account.</p><p>There is a large body of related work from the area of Information Retrieval that concerns the development of measures based on graded relevance as well as investigations of their properties <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b6">[7]</ref><ref type="bibr" target="#b7">[8]</ref><ref type="bibr" target="#b8">[9]</ref><ref type="bibr" target="#b9">[10]</ref><ref type="bibr" target="#b10">[11]</ref><ref type="bibr" target="#b11">[12]</ref>. We rely heavily on these achievements and our work can be viewed as an application and adaptation of this work to the SWS retrieval evaluation domain. We are not aware of any previous work on relevance schemes specifically designed for the SWS retrieval domain and discussions on how to provide reliable and consistent relevance judgments within this domain.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Relevance for SWS Retrieval Evaluation</head><p>The criteria most often used for experimental retrieval evaluation has been the effectiveness of a retrieval system, i.e. how good a system is in retrieving those and only those items that a user is interested in. Effectiveness evaluations are thus based on the notion of relevance of an item to a query <ref type="bibr" target="#b12">[13]</ref>. Most evaluation campaigns, in particular TREC <ref type="foot" target="#foot_1">3</ref> , have primarily been based on binary relevance, i.e. a document (in the terminology of TREC) was considered to be either relevant or irrelevant to a topic, but no further distinction was made.</p><p>The few attempts for quantitative SWS retrieval effectiveness have so far adopted this binary approach <ref type="bibr" target="#b13">[14]</ref><ref type="bibr" target="#b14">[15]</ref><ref type="bibr" target="#b15">[16]</ref><ref type="bibr" target="#b16">[17]</ref>. However, it has been argued that binary relevance is too coarse-grained to evaluate SWS matchmaking approaches <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b17">18]</ref>. This view is supported by the fact that nearly all SWS matchmaking algorithms are designed to support multiple degrees of match (DOMs). In a classic paper, Paolucci et al., for instance, proposed the use of exact, plug in, subsumes, and fail <ref type="bibr" target="#b18">[19]</ref>. This scale or variations thereof have been adopted by many approaches.</p><p>It is thus desireable to employ a graded relevance scale instead of a binary one in SWS retrieval evaluations. However, the design of such a scale is far from trivial.</p><p>To be practically useful it must have clear definitions that enable domain experts to provide reference relevance judgments as unambiguously as possible. In this aspect a scale like very relevant, relevant, somewhat relevant, slightly relevant, and irrelevant as used by <ref type="bibr" target="#b2">[3]</ref> is very difficult to judge objectively. On the other hand, human assessors should judge the relevance of a service offer with respect to a service request on the level of the original services and not their semantic formalizations. After all, the appropriateness and quality of these formalizations is also part of what is being evaluated. It is therefore not appropriate to directly use the DOMs by Paolucci et al. as a relevance scale for general SWS retrieval evaluation either. The definition of these DOMs is only meaningful in the context of DL subsumption reasoning, i.e. in the context of a particular formalization approach. It can not be meaningfully applied outside of this context.</p><p>To define a relevance scale that is equally applicable to different approaches but still sufficiently well defined to allow objective judgments, some assumptions and central terms need to be clarified. To this end, we recapitulate the basic definitions from a conceptual architecture for semantic web services presented by Preist <ref type="bibr" target="#b19">[20]</ref>. According to this architecture, a service is defined as a provision of value in some given domain, e.g. the booking of a particular flight ticket. Web services are technical means to access or provide such services. Service providers typically do not offer one particular service, but a set of coherent and logically related services, e.g. booking of flight tickets in general and not a specific flight ticket. Service descriptions will thus describe the set of services a provider is able to offer respectively a requester is interested in. Due to dynamics involved, privacy issues, and limited precision and detailedness, service descriptions will not always precisely capture the set of services that a provider is able to deliver or that a requester is interested in. Instead, they may be incorrect (not all described services can be provided or are of interest) as well as incomplete (not all services that can be provided or are of interest are covered by the description).</p><p>Keller et al. extended this model by remarking, that descriptions based on this model are not semantically unambiguous without knowing the intention of a modeler, which can be that either all or only some of the elements contained in the described service set are requested respectively can be delivered <ref type="bibr" target="#b20">[21]</ref>. Based upon this consideration they formally define different set theoretic match relationships between service offer and request descriptions. Because of its flexibility combined with clear definitions and its grounding to a well-defined conceptual model we propose a relevance scale that builds upon the match relationships introduced by Keller et al., extended by the notions of RelationMatch and Ex-cessMatch that we will explain below:</p><p>Match: The offer satisfies the request completely. PossMatch: The offer might satisfy the request completely, but due to the incompleteness of the descriptions this can not be guaranteed based upon the available information. ParMatch: The offer satisfies the request, but only partially (it offers some of the services which are requested but not all). PossParMatch: The offer might satisfy the request partially, but due to the incompleteness of the descriptions this cannot be guaranteed based upon the available information. RelationMatch: The offer does not provide services as requested, but related functionality. Thus, it could be useful in coordination with other offers. ExcessMatch: The offer is able to provide the requested services but would result in additional undesirable effects that are not requested by the client. NoMatch: None of the above, the offer is completely irrelevant to the request.</p><p>As a first remark, please note that these relevance degrees are not totally ordered. It will depend on the particular use case at hand, whether e.g. a definite partial match is preferable or not to a possible full match. Match, PossMatch, ParMatch, PossParMatch, and NoMatch have been introduced by Keller et al. We omit a detailed discussion due to space limitations and refer the interested reader to <ref type="bibr" target="#b20">[21]</ref>. Instead, we will focus on RelationMatch and ExcessMatch and motivate why these extensions are necessary.</p><p>A ParMatch characterizes a situation where the client requests multiple services and a provider is capable of delivering only some of those. A similar situation arises, when, for instance, a web service is able to deliver the desired effects, but the client is unable to provide the required inputs. Consider for instance a web service able to provide flight bookings between airports identified by the international airport code and a client that requests a flight between two particular cities. The web service can not be used directly to fulfill the client's request but intuitively it would still constitute a partial match. Such situations may arise in the context of all of the four typical elements of services: inputs, outputs, preconditions and effects. To distinguish such advertisements from completely irrelevant ones, but also from the clear defined ParMatch, we added the notion of RelationMatch.</p><p>We continue with a discussion of ExcessMatch. Typically, a full match between a service advertisement and request is defined as meeting the following conditions <ref type="bibr" target="#b1">[2]</ref>: All inputs required by the offer are available, the preconditions of the advertisement are satisfied by the state of the world prior to the service execution, and the offer provides all outputs and effects required by the client. The first two conditions concern the applicability of a service in a given situation, the last concerns its usefulness with respect to the client's request. Most approaches disregard a problem that arises, if a web service delivers more effects than are requested by the client. A client wanting to purchase a cell phone (only requested effect) would likely reject an advertisement that sells a cell phone (Effect 1) bundled with a contract with a specific telecommunication company (Effect 2). Nevertheless, most SWS matchmaking approaches would consider this a perfect match since all requested effects are delivered by the provider at hand. Similarly, a client looking for apartments in Berlin may or may not accept a web service providing a listing of apartment offers if that listing can not be restricted to offers located in Berlin. To accommodate such situations, we added the notion of ExcessMatch.</p><p>Finally, we would like to point out that, strictly spoken, the differentiations between Match and PossMatch (level of guarantee in the presence of imprecise descriptions), ParMatch and Match (level of horizontal completeness), Re-lationMatch and Match (issue of partial incompatibilities), and ExcessMatch and Match (issue of unwanted additional effects) are actually four unrelated dimensions that would result in 16 (2 4 ) levels of relevance even if each dimension is considered to be binary. To keep relevance levels manageable by the domain experts providing reference judgments, we restrict the scale to the seven relevance levels listed above for the time being. These seem to be the most important, but a further investigation of the optimal number of relevance levels is necessary and will be done in future work.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Evaluation Measures Based on Graded Relevance</head><p>To leverage the extra information contained in graded relevance judgments and graded degrees of match in a retrieval effectiveness evaluation, the retrieval measures for binary relevance need to be generalized to graded relevance. In this section, we present such generalized measures. To make the paper self-contained, we start by briefly recalling some basic definitions for the binary case.</p><p>Throughout this paper, we use the following definitions. Let R be the set of relevant items for a query. Let L be the set of items returned in response to that query. Then Recall is defined as the proportion of relevant items returned (Recall = L R R ) and Precision as the proportion of returned items that are relevant (P recision = L R L ). Recall and Precision are set-based measures. However, there is an obvious trade-off between them. By returning more items, a system can usually increase its Recall at the expense of its Precision. Thus, in the following we assume that systems return a ranked output ordered by estimated confidence in relevance. Let r, 1 &lt;= r &lt;= |L| denote a specific rank in this output. Let isrel(r) = 1, if the item at rank r is relevant and 0 otherwise. Let count(r) be the number of relevant items among the top r retrieved items, i.e. count(r) = r i=1 isrel(i). This allows to measure Precision as a function of Recall by scanning L from the top to the bottom and measure the Precision at standard Recall levels. These measures average well for different queries and the corresponding R/P charts are the most widely used measure to compare the retrieval performance of systems. It is also possible to measure Precision and Recall at predefined ranks (P recision r and Recall r , r is often referred to as document cutoff level). However, these measures do not average well for queries where |R| varies greatly.</p><p>If a system's performance needs to be captured in a single measure, the probably most often used one is Average Precision over relevant items which is defined as:</p><formula xml:id="formula_0">AveP = 1 |R| |L| r=1 isrel(r) count(r) r .</formula><p>Since about 2000, there is an increased interest in measures based on graded or continous relevance. Various proposals have been made to generalize the measures introduced above from binary to graded relevance (see <ref type="bibr" target="#b11">[12]</ref> for a discussion). Most of these are based on or can be expressed in terms of Cumulated Gain proposed by Järvelin and Kekäläinen <ref type="bibr" target="#b7">[8]</ref>. Intuitively, Cumulated Gain at rank r measures the gain that a user receives by scanning the top r items in a ranked output list. More formally let g(r) &gt;= 0 denote the gain value (or the relevance level) of the item at rank r and from now on isrel(r) = 1, if g(r) &gt; 0 and 0 otherwise. Then Cumulated Gain at rank r is defined as cg(r) = r i=0 g(r). Moreover consider an ideal ranking, i.e. ∀(r &gt; 1, r &lt;= |R|) : isrel(r) = 1 and ∀(r &gt; 1) : g(r) &lt;= g(r − 1). Let icg(r) (Ideal Cumulated Gain at rank r) denote the Cumulated Gain for this ideal ranking.</p><p>Since cg(r) can take arbitrarily large values for queries with many relevant items it has to be normalized to average or compare results across queries. Normalized Cumulated Gain<ref type="foot" target="#foot_2">4</ref> at rank r is defined as the retrieval performance relative to the optimal retrieval behavior, i.e. ncg(r) = cg (r)  icg(r) . It allows a straightforward extension of AveP which has sometimes been referred to as Average Weighted Precision <ref type="bibr" target="#b4">[5]</ref>: cg(r)  icg(r) . Unfortunately, ncg(r) has a significant flaw that AWP inherits: since icg(r) has a fixed upper bound (icg(r) &lt;= icg(|R|)), ncg(r) and AWP cannot penalize late retrieval of relevant documents properly since ncg(r) cannot distinguish at which rank relevant documents are retrieved for ranks greater than R <ref type="bibr" target="#b10">[11]</ref>. This can be illustrated by comparing ncg(r) and P recision r for the last rank in a full output (R ⊆ L). In this case ncg(|L|) = 1 but P recision(|L|) = |R| |L| , which is usually much smaller than one. Several measures have been proposed that resolve this flaw of AWP.</p><formula xml:id="formula_1">AW P = 1 |R| |L| r=1 isrel(r)</formula><p>Järvelin and Kekäläinen <ref type="bibr" target="#b7">[8]</ref> suggested to use a discount factor to penalize late retrieval and thus reward systems that retrieve highly relevant items early. They defined Discounted Cumulated Gain at rank r as dcg(r) = r i=0 g(r) disc(r) with disc(r) &gt;= 1 being an appropriate discount function. Järvelin and Kekäläinen suggest to use the log function and use its base b to customize the discount which leads to disc(r) = log b r for r &gt; b and disc(r) = 1 otherwise (the distinction is necessary to maintain disc(r) &gt;= 1 to avoid boosting the first ranks).</p><p>We use an according definition of Ideal Discounted Cumulated Gain (idcg(r)) to define an adapted Version of AWP that we call Average Weighted Discounted Precision:</p><formula xml:id="formula_2">AW DP = 1 |R| |L| r=1 isrel(</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>r) dcg(r) idcg(r) .</head><p>Similarly, Kishida <ref type="bibr" target="#b11">[12]</ref> proposed a generalization of AveP that also avoids the flaw of AWP:</p><formula xml:id="formula_3">genAveP = |L| r=1 isrel(r) cg(r) r |R| r=1 icg(r) r</formula><p>Furthermore, Sakai <ref type="bibr" target="#b4">[5]</ref> proposed an integration of AWP and AveP called Qmeasure which inherits properties of both measures and possesses a parameter β to control whether Q-measure behaves more like AWP or more like AveP:</p><formula xml:id="formula_4">Q-measure = 1 |R| |L| r=1 isrel(r) βcg(r) + count(r) βicg(r) + r</formula><p>All measures allow to finetune the extent to which highly relevant items are preferred over less relevant items (by setting appropriate gain values) but differ in the degree of control that is possible with respect to the extent to which late retrieval is penalized. Q-Measure controls the penalty by its β Parameter, AW DP by the choice of an appropriate discounting function, and genAveP lacks such control. Sakai <ref type="bibr" target="#b8">[9]</ref> discusses this issue in detail but unfortunately disregards choices of disounting functions for ndcg(r) besides the logarithm.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Experimental Retrieval Evaluation</head><p>We now report on the evaluation of our appraoch by means of a preliminary experiment on using the relevance scale introduced in Section 3 and the measures introduced in the previous section to evaluate the retrieval effectiveness of two matchmakers. We start by describing the test data we used and in particular our experiences on obtaining graded relevance judgments. We continue by describing the parameters that we chose for the experiment and complete our report with a discussion of our results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1">Test Data</head><p>Unfortunately, there is still a lack of standard test collections in the area of SWS <ref type="bibr" target="#b17">[18]</ref>. To test the proposed evaluation approach, we chose the Education subset of the OWLS-TC 2.2 test collection <ref type="foot" target="#foot_3">5</ref> . This subset contains 276 OWL-S service descriptions and six request descriptions together with binary relevance judgments. We chose this subset mainly for two reasons. First, this subset<ref type="foot" target="#foot_4">6</ref> had been used previously in an experiment with graded relevance judgments which allows to compare our results with the results from that previous experiment <ref type="bibr" target="#b2">[3]</ref>. Second, for OWLS-TC, ranked outputs from two different matchmakers, OWLS-M3 <ref type="bibr" target="#b13">[14]</ref> and iMatcher <ref type="bibr" target="#b15">[16]</ref>, are available through the organizers of the S3 Matchmaker Contest<ref type="foot" target="#foot_5">7</ref> . However, it turned out that iMatcher was unable to process one of the six queries which was thus excluded from the test data. Further information including all test data and results are available online <ref type="foot" target="#foot_6">8</ref> .</p><p>To collect and manage graded relevance judgments for this subset, we used the OPOSSum portal <ref type="foot" target="#foot_7">9</ref> which already lists all the OWLS-TC services. Therefore, throughout this paper we identify queries by their id from that portal (5654, 5659, 5664, 5668, and 5675). We extended OPOSSum with a user interface that allows to conveniently enter graded relevance judgments for large numbers of services. We developed some guidelines for relevance judges <ref type="foot" target="#foot_8">10</ref> and three persons (one expert in the area of SWS as well as two volunteers that had only a basic understanding of SWS) judged the complete subset.</p><p>Unfortunately, it turned out that the judgments of the three judges did not correspond with each other very well. We believe that this is largely caused by insufficient textual documentation of the services in the employed test collection. This lack of detail required relevance judges to make a lot of assumptions regarding the semantics of the services. Consequently, single judges were able to judge consistently but judgments varied between the judges depending on the different assumptions that were made (for instance whether a lecturer or a research assistant are considered researchers or not). For the rest of this paper and the reported preliminary experiment we used the judgments of the SWS expert exclusively.</p><p>We compared these judgments with the binary OWLS-TC judgments. Table <ref type="table" target="#tab_0">1</ref> shows that correspondance. For each graded relevance level it shows how many of the services judged into this level were evaluated relevant versus irrelevant by the OWLS-TC authors. The average row shows the arithmetic mean that is computed by assigning a value of one/zero to the binary relevant/irrelevant services. Please note that none of the services in the Education subset of OWLS-TC was judged an ExcessMatch by our judges. Nevertheless we believe that this relevance level has its own right of existence for other collections.</p><p>Since OWLS-TC employs a very liberal definition of relevance, we were surprised to see eight services judged irrelevant by OWLS-TC but judged a perfect Match by our judges. A closer look revealed that seven of those eight mismatches seem to indicate errors in the OWLS-TC reference judgments. The remaining mismatch is caused by different context knowledge assumptions. Such assumptions also explain most of the other mismatches, like the twenty services judged irrelevant by us but relevant by OWLS-TC. Most of these, for instance, are related to a query for scholarships. Services providing information about loans were judged relevant by OWLS-TC but irrelevant by our expert.</p><p>Finally, we compared our judgments with the fuzzy relevance judgments made by Tsetsos and colleagues <ref type="bibr" target="#b2">[3]</ref> for the OWLS-TC 2.1 Education subset, which contains the same requests as the 2.2 subset but only 135 instead of 276 services. Tsetsos et al. used a fuzzy scale with the values irrelevant, slightly rele-vant, somewhat relevant, relevant, and very relevant. For each graded relevance level Table <ref type="table" target="#tab_1">2</ref> shows how many of the services judged into this level by our judges were judged into each of their fuzzy levels by Tsetsos et al. The average values are computed by assigning values of zero through four to the relevance levels used by Tsetsos et al. The small numbers in the Irrelevant row are caused by the fact that we used only explicit judgments, but Tsetsos et al. provided most "irrelevant" judgments only implicitly. Thus, with a full set of explicit judgments, numbers in the Irrelevant row would have been much higher and the Averages in particular in the last column much lower. We were surprised that the services judged as a perfect Match by our judges were relatively evenly distributed among the four top relevance levels of Tsetsos et al. (see first column). Since we could not obtain information about the rationale behind those judgments or the precise definitions of the relevance levels we lack an explanation for this phenomena but we expect it to be caused by the same issues that caused our judges to judge differently relatively often, too.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2">Evaluation Parameters</head><p>The measures described in Section 4 allow to evaluate SWS retrieval systems based on the graded relevance scheme introduced in Section 3 but leave open the question about the proper parameter combinations to use in an evaluation. As Järvelin and Kekäläinen remark, "the mathematics work for whatever parameter combinations and cannot advise us on which to choose. Such advise must come from the evaluation context in the form of realistic evaluation scenarios" <ref type="bibr" target="#b7">[8]</ref>. In order to perform an investigation in particular of the effects of switching from binary to graded relevance, we chose two gain value settings that actually correspond to binary relevance and two settings which leverage the potential of graded relevance. The corresponding gain values are displayed in Table <ref type="table" target="#tab_2">3</ref>. Strict Binary and Relaxed Binary correspond to strict versus relaxed definitions of binary relevance. Graded 1 corresponds to a setting with a focus on high precision which is appropriate in a use case of automated dynamic binding whereas Graded 2 reflects a more balanced preference between precision and recall which seems more appropriate in use cases where a human programmer is searching for a service. Additionally (not shown in Table <ref type="table" target="#tab_2">3</ref>) we used the binary relevance judgments that come together with OWLS-TC 2.2 for comparison.</p><p>For each of the five queries and each of the five gain value settings, we computed the following measures for both matchmakers: AWDP using the discount functions r (AWDP-R), √ r (AWDPSQRT), log 2 r (AWDPLog2) and log 10 r (AWDPLog10) as well as without discount function (AWP), Q-Measure with β = 5, β = 1, β = 0.5, and β = 0, and genAveP. Using a quickly growing discount function in conjunction with AWDP (e.g. AWDP-R) rewards systems that retrieve highly relevant items early, i.e. it puts the emphasis of the evaluation on the top ranks. Using no discount function (AWP) leads to a more balanced consideration of all ranks at the prize of loosing the ability to penalize a very late retrieval of items. Slowly growing discount functions (e.g. AWDPLog2) constitute a compromise between these extremes. In the case of Q-Measure a larger β makes Q-Measure more similar to AWP, i.e. rewards retrieving highly relevant items prior to marginally relevant items but makes it vulnerable to not penalizing very late retrieval of relevant items. Small choices for β make Q-Measure more similar to the traditional binary AveP which does not differentiate between highly and marginally relevant items but correctly penalizes late retrieval of relevant items. A choice of β = 0 completely reduces Q-Measure to binary AveP. Similarly, genAveP is reduced to AveP in settings with binary relevance.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3">Results</head><p>As expected, results vary significantly over queries. For Query 5675, for instance, M3 is rated higher by 40 out of the 50 possible combinations of measures and gain value settings. In contrast, for Query 5675 iMatcher is rated higher by all measures. Given this large variation, the relatively small size of our data set and in particular the fact that we had data only from two matchmakers, the results that we report in the remainder of this section need to be taken with a grain of salt. Nevertheless, they indicate some interesting preliminary findings.</p><p>Our results confirm the expectation, that the choice of measure matters, not only in terms of absolute numbers but also in terms of which matchmaker is rated higher. This is illustrated by Figure <ref type="figure" target="#fig_0">1</ref> that shows the values of the various measures for Request 5654 and Strict Binary and Graded 1 gain value settings. For this request, AWDP with large discounting favors iMatcher while AWDP with little or no discounting as well as Q-measure favor M3. 11  While results frequently changed with different measures, we found that, except for β = 0, the choice of β has little influence on the absolute and relative performance of the matchmakers (see Figure <ref type="figure" target="#fig_0">1</ref>). In fact, with our data, different parameterizations for Q-measure almost never made a difference in terms of which matcher is rated higher. Furthermore, genAveP always ranked the two matchmakers the same way Q-measure did.</p><p>For the binary cases, this behavior of Q-measure can be well explained. In this case cg(r) = count(r) and icg(r) = r if r &lt;= |R|. Thus, Q-measures fraction can 11 Please note that this finding (Q-measure favors M3) are specific to this request. We found frequent changes of the favored matchmaker when changing the measure but no general trend that a particular measure favors a particular matchmaker. be reduced by β + 1 if r &lt;= |R|. In other words, in binary cases, β influences the value of Q-measure only for relevant items retrieved after rank |R|. Relatively few relevant items are retrieved at such ranks in our experiment, thus the influence of β to the value of Q-measure is limited with our data. Compared to the influence of Q-measure's β, the choice of discount function of AWDP caused more changes in the ratings. It didn't cause changes in the ratings for Queries 5664, 5668, and 5675 but for the remaining two queries the different versions of AWDP disagreed in eight of ten cases (two queries times five gain value settings), including those displayed in Figure <ref type="figure" target="#fig_0">1</ref>.</p><p>The notable peak of both matchmaker's performance for the Graded 1 gain value setting measured with Q0 that is visible in Figure <ref type="figure" target="#fig_0">1</ref> highlights how the use of graded relevance influences measure results. Using β = 0 reduces Q-measure to AveP and thus Graded 1 to a binary scale which largely resembles the original OWLS-TC judgments. For both matchmakers, this results in a significantly increased absolute performance, albeit not in a change of their performance relative to each other.</p><p>Generally, changes in the settings of the gain values caused more significant changes in how the matchmakers were rated than changes in the parameterizations of AWDP and Q-measure. However, the Q-measure variations and genAveP were again less sensitive towards changes in the evaluation parameters than the AWDP-family. Their ratings did not change regardless of the gain values used except for Query 5664 where they all preferred M3 with the Strict Binary and iMatcher with the other settings 12 . In contrast, with the one exception of Query 5668 where iMatcher outperforms M3 regardless of the measure, changes in the gain value settings caused changes in the ratings of the AWDP measure family in nearly half of the cases. As an example, Figure <ref type="figure" target="#fig_1">2</ref> shows the values of AWD-PLog2 and AWP for Request 5664: M3 is favored by both measures for Strict Binary and Graded 1 while iMeasure is favored for Relaxed Binary, OWLS-TC Binary, and Graded 2. Generally, Relaxed Binary and OWLS-TC Binary tend to benefit iMatcher while the other settings tend to benefit M3. The most likely interpretation is that M3 performs a stricter selection and thus outperforms iMatcher in ranking more relevant services higher. Another influence factor may be that iMatcher applies machine learning techniques and has been trained with the binary relevance judgments of OWLS-TC. Switching to other definitions of relevance (e.g. strict binary relevance) seems to have a negative impact on iMatcher's performance relative to that of M3.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Conclusions and Future Work</head><p>In this paper, we investigated how to evaluate the retrieval effectiveness of SWS matchmakers based on graded instead of binary relevance. We discussed the notion of relevance in this particular context and proposed a well-founded scale of relevance levels customized to the SWS matchmaking domain. We presented a number of evaluation measures for graded relevance and described an experiment of using those measures to perform a retrieval evaluation of two SWS matchmakers.</p><p>We need to note once more, that our results have to be considered preliminary because of the nature of the test data used. First, there was a significant variation in how the different judges judged our test data. We believe this to be caused by the insufficient documentation of the services in our data and expect this issue to improve if more realistic and better documented services are used than are currently available in the form of OWLS-TC. Second, we have compared only two matchmakers based on a relatively small data set. In terms of investigating the effects of different measures and different relevance scales it would be particularly desireable to have access to a larger number of directly comparable matchmakers. This will be the case if either more readily implemented matchmakers for a particular formalism (for instance SAWSDL) become available or test collections across formalisms will be developed.</p><p>Despite of these two restrictions, our results allow to draw a number of interesting conclusions. First, retrieval evaluation based on graded relevance is feasible both in terms of the effort to obtain graded instead of binary relevance judgments and in terms of the availability of measures suitable for graded relevance. Second, the choice of gain values (i.e. relevance levels) and the choice of measure has a significant influence on the evaluation results. Our results indicate that the choice of gain values has a greater impact than the choice of measure. Third, AWDP seems to be more sensitive towards changes in the parameterizations (regarding both, the penalty for late retrieval and changes of the gain values) than Q-measure and thus should probably be the first choice for future evaluations.</p><p>In our future work we plan to verify these findings with better data. As a first step, we would like to investigate whether relevance judgments really become more consistent across judges when more realistic and well documented services are used. A second step will then be to compare a larger number of matchmakers based on that more realistic data.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Results for Request 5654 with Strict Binary and Graded 1 gain values.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. AWDP measures for Request 5664 for different gain values.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 .</head><label>1</label><figDesc>Correspondance with binary OWLS-TC 2.2 judgments</figDesc><table><row><cell cols="2">Match</cell><cell>Poss</cell><cell>Par</cell><cell cols="3">PossPar Relation Excess</cell><cell>None</cell></row><row><cell cols="2">Relevant 130</cell><cell>12</cell><cell>33</cell><cell>5</cell><cell>6</cell><cell>-</cell><cell>20</cell></row><row><cell>Irrelevant 8</cell><cell></cell><cell>3</cell><cell>7</cell><cell>1</cell><cell>37</cell><cell>-</cell><cell>1408</cell></row><row><cell cols="2">Average 0.94</cell><cell>0.8</cell><cell>0.83</cell><cell>0.83</cell><cell>0.14</cell><cell>-</cell><cell>0.01</cell></row><row><cell></cell><cell>Match</cell><cell>Poss</cell><cell>Par</cell><cell cols="3">PossPar Relation Excess</cell><cell>None</cell></row><row><cell>Very r.</cell><cell>24</cell><cell>1</cell><cell>4</cell><cell>0</cell><cell>0</cell><cell>-</cell><cell>0</cell></row><row><cell>Relevant</cell><cell>19</cell><cell>1</cell><cell>2</cell><cell>0</cell><cell>0</cell><cell>-</cell><cell>2</cell></row><row><cell>Slightly r.</cell><cell>11</cell><cell>7</cell><cell>1</cell><cell>0</cell><cell>1</cell><cell>-</cell><cell>1</cell></row><row><cell cols="2">Somewhat r. 10</cell><cell>2</cell><cell>3</cell><cell>2</cell><cell>2</cell><cell>-</cell><cell>4</cell></row><row><cell>Irrelevant</cell><cell>3</cell><cell>0</cell><cell>0</cell><cell>1</cell><cell>0</cell><cell>-</cell><cell>15</cell></row><row><cell>Average</cell><cell>2.75</cell><cell>1.64</cell><cell>2.9</cell><cell>1.33</cell><cell>1.5</cell><cell>-</cell><cell>0.68</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 .</head><label>2</label><figDesc>Correspondance with fuzzy judgments by Tsetsos et al.</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3 .</head><label>3</label><figDesc>Experimental gain value settings</figDesc><table><row><cell></cell><cell>Strict Binary</cell><cell cols="2">Relaxed Binary Graded 1</cell><cell>Graded 2</cell></row><row><cell>Match</cell><cell>1</cell><cell>1</cell><cell>6</cell><cell>4</cell></row><row><cell>PossMatch</cell><cell>0</cell><cell>1</cell><cell>2</cell><cell>2</cell></row><row><cell>ParMatch</cell><cell>0</cell><cell>1</cell><cell>1</cell><cell>2</cell></row><row><cell cols="2">PossParMatch 0</cell><cell>1</cell><cell>0.5</cell><cell>1</cell></row><row><cell cols="2">RelationMatch 0</cell><cell>1</cell><cell>0</cell><cell>2</cell></row><row><cell>ExcessMatch</cell><cell>0</cell><cell>1</cell><cell>0</cell><cell>1</cell></row><row><cell>NoMatch</cell><cell>0</cell><cell>0</cell><cell>0</cell><cell>0</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_0">http://www-ags.dfki.uni-sb.de/∼klusch/s3/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_1">http://trec.nist.gov/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_2">A similar measure has been proposed by Pollack in 1968 under the name sliding ratio.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_3">http://projects.semwebcentral.org/projects/owls-tc/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_4">More precisely a similar subset from a smaller previous release of this test collection.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_5">http://www-ags.dfki.uni-sb.de/∼klusch/s3/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_6">http://fusion.cs.uni-jena.de/OPOSSum/ISWC08-SMRR/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_7">http://fusion.cs.uni-jena.de/OPOSSum</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="10" xml:id="foot_8">http://fusion.cs.uni-jena.de/OPOSSum/index.php?action=relevanceguideline</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>We would like to thank Patrick Kapahnke and Matthias Klusch for providing us with test data from the S3 Contest and for their help in resolving some technical problems with that data. We would also like to thank them and the other contributors for developing OWLS-TC and making it publicly available. Finally, we would like to thank Vassileios Tsetsos for giving us the graded relevance judgments from <ref type="bibr" target="#b2">[3]</ref> to compare them with ours.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Semantic web services</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">A</forename><surname>Mcilraith</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">C</forename><surname>Son</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Zeng</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Intelligent Systems</title>
		<imprint>
			<biblScope unit="volume">16</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="46" to="53" />
			<date type="published" when="2001">2001</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Semantic web service coordination</title>
		<author>
			<persName><forename type="first">M</forename><surname>Klusch</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CASCOM -Intelligent Service Coordination in the Semantic Web</title>
				<editor>
			<persName><forename type="first">M</forename><surname>Schumacher</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><forename type="middle">H</forename></persName>
		</editor>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">On the evaluation of semantic web service matchmaking systems</title>
		<author>
			<persName><forename type="first">V</forename><surname>Tsetsos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Anagnostopoulos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Hadjiefthymiades</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">4th IEEE European Conference on Web Services (ECOWS2006)</title>
				<meeting><address><addrLine>Zürich, Switzerland</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Measuring retrieval effectiveness with average distance measure (ADM)</title>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">D</forename><surname>Mea</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Demartini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">D</forename><surname>Gaspero</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Mizzaro</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Information Wissenschaft und Praxis</title>
		<imprint>
			<biblScope unit="volume">57</biblScope>
			<biblScope unit="issue">8</biblScope>
			<biblScope unit="page" from="405" to="416" />
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">New performance metrics based on multigrade relevance: Their application to question answering</title>
		<author>
			<persName><forename type="first">T</forename><surname>Sakai</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Fourth NTCIR Workshop on Research in Information Access Technologies, Information Retrieval, Question Answering and Summarization (NTCIR04)</title>
				<meeting><address><addrLine>Tokyo, Japan</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Semantic matchmaking as non-monotonic reasoning: A description logic approach</title>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">D</forename><surname>Noia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">D</forename><surname>Sciascio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">M</forename><surname>Donini</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Artificial Intelligence Research (JAIR)</title>
		<imprint>
			<biblScope unit="volume">29</biblScope>
			<biblScope unit="page" from="269" to="307" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Using graded relevance assessments in IR evaluation</title>
		<author>
			<persName><forename type="first">Järvelin</forename><surname>Kekäläinen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of the American Society for Information Science and Technology</title>
		<imprint>
			<biblScope unit="volume">53</biblScope>
			<biblScope unit="issue">13</biblScope>
			<biblScope unit="page" from="1120" to="1129" />
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Cumulated gain-based evaluation of ir techniques</title>
		<author>
			<persName><forename type="first">K</forename><surname>Järvelin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kekäläinen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM Transactions on Information Systems</title>
		<imprint>
			<biblScope unit="volume">20</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="422" to="446" />
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">On penalising late arrival of relevant documents in information retrieval evaluation with graded relevance</title>
		<author>
			<persName><forename type="first">T</forename><surname>Sakai</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">First International Workshop on Evaluating Information Access (EVIA)</title>
				<meeting><address><addrLine>Tokyo, Japan</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">On the reliability of information retrieval metrics based on graded relevance</title>
		<author>
			<persName><forename type="first">T</forename><surname>Sakai</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Information Processing and Management</title>
		<imprint>
			<biblScope unit="volume">43</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="531" to="548" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Ranking the NTCIR systems based on multigrade relevance</title>
		<author>
			<persName><forename type="first">T</forename><surname>Sakai</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Revised Selected Papers of the Asia Information Retrieval Symposium</title>
				<meeting><address><addrLine>Beijing, China</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2004">2004</date>
			<biblScope unit="page" from="251" to="262" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">Property of average precision and its generalization: An examination of evaluation indicator for information retrieval experiments</title>
		<author>
			<persName><forename type="first">K</forename><surname>Kishida</surname></persName>
		</author>
		<idno>NII-2005-014E</idno>
		<imprint>
			<date type="published" when="2005">2005</date>
			<pubPlace>Tokyo, Japan</pubPlace>
		</imprint>
		<respStmt>
			<orgName>National Institute of Informatics</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Technical Report</note>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Relevance: The whole history</title>
		<author>
			<persName><forename type="first">S</forename><surname>Mizzaro</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">JASIS</title>
		<imprint>
			<biblScope unit="volume">48</biblScope>
			<biblScope unit="issue">9</biblScope>
			<biblScope unit="page" from="810" to="832" />
			<date type="published" when="1997">1997</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Automated semantic web service discovery with OWLS-MX</title>
		<author>
			<persName><forename type="first">M</forename><surname>Klusch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Fries</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Sycara</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">5th International Joint Conference on Autonomous Agents and Multiagent Systems (AAMAS 2006)</title>
				<meeting><address><addrLine>Hakodate, Japan</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Performance of hybrid WSML service matching with WSMO-MX: Preliminary results</title>
		<author>
			<persName><forename type="first">F</forename><surname>Kaufer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Klusch</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">First International Joint Workshop SMR 2 on Service Matchmaking and Resource Retrieval in the Semantic Web at the 6th International Semantic Web Conference (ISWC2007)</title>
				<meeting><address><addrLine>Busan, South Korea</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">The creation and evaluation of iSPARQL strategies for matchmaking</title>
		<author>
			<persName><forename type="first">C</forename><surname>Kiefer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bernstein</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">5th European Semantic Web Conference (ESWC2008)</title>
				<meeting><address><addrLine>Tenerife, Canary Islands, Spain</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="463" to="477" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><surname>Dudev</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Kapahnke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Klusch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Misutka</surname></persName>
		</author>
		<ptr target="http://www-ags.dfki.uni-sb.de/∼klusch/s3/s3-contest-plugin.zip" />
		<title level="m">International semantic service selection contest s3 -contest participation guideline</title>
				<imprint>
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">On the empirical evaluation of semantic web service approaches: Towards common SWS test collections</title>
		<author>
			<persName><forename type="first">U</forename><surname>Küster</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>König-Ries</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">2nd IEEE International Conference on Semantic Computing (ICSC2008)</title>
				<meeting><address><addrLine>Santa Clara, CA, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Semantic matching of web services capabilities</title>
		<author>
			<persName><forename type="first">M</forename><surname>Paolucci</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Kawamura</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">R</forename><surname>Payne</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">P</forename><surname>Sycara</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">First International Semantic Web Conference (ISWC2002)</title>
				<meeting><address><addrLine>Sardinia, Italy</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2002">2002</date>
			<biblScope unit="page" from="333" to="347" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<title level="m" type="main">A conceptual architecture for semantic web services (extended version)</title>
		<author>
			<persName><forename type="first">C</forename><surname>Preist</surname></persName>
		</author>
		<idno>HPL-2004-215</idno>
		<imprint>
			<date type="published" when="2004">2004</date>
		</imprint>
		<respStmt>
			<orgName>HP Laboratories Bristol</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Technical Report</note>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Automatic location of services</title>
		<author>
			<persName><forename type="first">U</forename><surname>Keller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Lara</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Lausen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Polleres</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Fensel</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Second European Semantic Web Conference (ESWC2005)</title>
				<meeting><address><addrLine>Heraklion, Crete, Greece</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
