<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Checklist-Based Approach for Quality Assessment of Scienti c Information</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Jun Zhao</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Graham Klyne</string-name>
          <email>graham.klyne@zoo.ox.ac.uk</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Matthew Gamble</string-name>
          <email>m.gamble@cs.man.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Carole Goble</string-name>
          <email>carole.goble@manchester.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Computer Science, University of Manchester</institution>
          ,
          <addr-line>Manchester, M13 9PL</addr-line>
          ,
          <country country="UK">UK</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of Zoology, University of Oxford</institution>
          ,
          <addr-line>Oxford, OX1 3PS</addr-line>
          ,
          <country>UK jun.zhao</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The Semantic Web is becoming a major platform for disseminating and sharing scienti c data and results. Quality of these information is a critical factor in selecting and reusing them. Existing quality assessment approaches in the Semantic Web largely focus on using general quality dimensions (accuracy, relevancy, etc.) to establish quality metrics. However, speci c quality assessment tasks may not t into these dimensions and scientists may nd these dimensions too general for expressing their speci c needs. Therefore, we present a checklist-based approach, which allows the expression of speci c quality requirements, saving users from the constraints of the existing quality dimensions. We demonstrate our approach by two scenarios and share our lessons about di erent semantic web technologies that were tested during our implementation.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Information quality assessment aims to provide an indication of the tness of
information. Most existing approaches perform the assessment by integrating
assessment of a number of quality dimensions, such as accuracy, completeness,
or believability. We argue that although such methodology provides a systematic
framework to organise quality assessment, it leaves two outstanding issues: 1) the
quality dimensions used are often too abstract and generic for expressing concrete
quality requirements, and 2) constrained frameworks are often unable to address
di erent uses a consumer may have for a common resource: data t for one
purpose might not be t for another. Although quality dimensions are often
specialised to support assessment requirements from a speci c domain or task,
e.g. as a formula to compute a quality value by using a certain set of information,
such specialisation cannot always be exible enough to support di erent quality
needs that might arise from di erent tasks to be applied to the same information.
For example, the set of information considered su cient for supporting access
to a linked data resource might not be enough for assessing its freshness. Users
need a exible way to express their di erent quality requirements according to
the task at hand.</p>
      <p>
        This paper addresses these issues by proposing a exible and extensible data
model to support explicit expression of quality requirements. We draw upon
the idea of checklists, a well-established tool for ensuring safety, quality and
consistency in complex operations, such as manufacturing or critical care [
        <xref ref-type="bibr" rid="ref4">4,
10</xref>
        ]. A checklist explicitly de nes a list of requirements that must be ful lled
or assessed for a given task. In our checklist-based framework we provide an
OWL ontology, the Minim ontology, to express quality requirements as RDF,
and an assessment tool to evaluate the conformance of target data against a
Minim checklist. We demonstrate Minim in practice by applying it to support
two quality assessment scenarios: the quality of scienti c data, and scholarly
artefacts.
      </p>
      <p>The contributions of this paper are: 1) presenting a exible and extensible
data model for explicitly expressing quality requirements according to users'
assessment needs; and 2) providing a comparison of several state-of-the-art
semantic web technologies in supporting quality assessment tasks, which are learnt
from our practical experience. The Minim model presented in this work is an
updated version of our previous work [14], which provide two new distinct
features: 1) more explicit representation of individual quality requirement as a type
of test; and 2) an extensible structure for users to add requirements or tests that
are not de ned in the model, in order to cope with new emerging requirements
from their own domains.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Motivating Scenarios</title>
      <p>In this section we present our motivating quality assessment scenarios from the
scienti c and scholarly publishing domains. The scenarios illustrate how our
checklist framework can be used to support speci c quality assessment tasks.
Although these requirements could be t into a conventional quality dimension,
such as correctness or completeness, our approach saved the users from having
to take the extra step of identifying the relevant quality dimensions, which is
commonly required in an existing dimension-based methodology. Therefore, our
scenarios highlight the advantage and convenience of being able to explicitly
express the assessment requirements using our approach.
2.1</p>
      <sec id="sec-2-1">
        <title>Quality assessment of scienti c linked data</title>
        <p>The volume of scienti c data resources on the linked data web is rapidly
expanding. However, their quality does not always stand up to scrutiny, an issue that is
caused either by the linked data publication process or is intrinsic to the source
data. Scenario 1 shows how quality assessment can reveal a series of potential
quality issues in a linked dataset that contains some basic metadata information
about 7,572 chemical compounds. The dataset was used in a previous study [7]
and it was created based on the InfoBox information of Wikipedia3. Because of</p>
        <sec id="sec-2-1-1">
          <title>3 http://en.wikipedia.org/</title>
          <p>the potential incompleteness of the information available from these InfoBoxes,
the resulting linked dataset can also have some potential quality issues. For
example, according to domain-speci c recommendations, each chemical compound
must have one and only one IUPAC International Chemical Identi er (InChI).
A quality requirement like this can be easily expressed using the cardinality
test construct in our checklist model (see section 3) and an assessment can be
automatically performed against all the chemical compounds in the dataset.
2.2</p>
        </sec>
      </sec>
      <sec id="sec-2-2">
        <title>Quality assessment for scholarly communication</title>
        <p>Scholarly communication refers to a principled method of making scienti c
artefacts available in order to support their more e ective interpretation and reuse.
These artefacts include data, methods or tools that were used to generate the
ndings reported, and providing su cient information is key to achieving this
goal. This is an ongoing quality challenge in scholarly communication that has
not been fully addressed.</p>
        <p>Scenario 2 uses quality assessment to help boost the e ectiveness of
scholarly communication in practice. myExperiment.org [5] is a popular work ow
repository for sharing and releasing scienti c work ows, which are important
rst-class scienti c artefacts documenting protocols used to generate
experimental results. Re-use of these work ows relies on adequate documentation to
facilitate understanding and re-purposing.</p>
        <p>A previous study analysed a representative selection work ows from
myExperiment.org and drew out a minimal set of information that supports their
re-execution [14]. This information, presented as a quality checklist, can be used
to prompt work ow authors to provide better documentation about the
workows. This early intervention enhances the quality of scholarly communication.
2.3</p>
      </sec>
      <sec id="sec-2-3">
        <title>Summary</title>
        <p>No quality dimensions need be mentioned in the quality requirements of our
scenarios. Instead, these requirements can be directly expressed using the
constructs of our checklist data model, see sections 3 and 6. This provides a novel
approach to quality assessment, in comparison to most of the existing work.
3</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Approach</title>
      <p>Our checklist-based assessment approach is based on two central pieces: 1) a
container data model for encapsulating the RDF data/graph to be evaluated, and
2) the Minim data model, for representing quality requirements as a checklist.
3.1</p>
      <sec id="sec-3-1">
        <title>Research Object Model as a Container</title>
        <p>
          We use an existing data model, namely the Research Object (RO) model [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ],
for our assessment. This provides a lightweight `container' structure for
encapsulating RDF and associated data. Annotation data contained within the RO
constitutes the collection of RDF descriptions to be evaluated.
3.2
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>The Minim Model for Expressing Quality Requirements</title>
        <p>A checklist provides an overall assessment of a dataset for some purpose. It
consists of a number of individual checklist items which may address speci c
values within a dataset (typically at the level granularity accessible by a SPARQL
query). Borrowing from IETF practice 4, individual items have a MUST, SHOULD
or MAY requirement level. A dataset may be \fully compliant", \nominally
compliant" or \minimally compliant" with a checklist if it satis es all of its MAY,
SHOULD or MUST items respectively.</p>
        <p>Our Minim data model (see Figure 1) provides 4 core constructs to express
a quality requirement:
{ minim:Checklist5, to associate a RO context, a target (the RO or a resource
within the RO) and an assessment purpose (e.g. runnable work ow) with a
minim:Model to be evaluated.
{ minim:Model, to enumerate the requirements (checklist items) to be
evaluated, with corresponding MUST, SHOULD or MAY requirement levels.
{ minim:Requirement, which is a single requirement (checklist item) that is
associated with a minim:Rule for evaluating whether or not it is satis ed or
not satis ed.
{ minim:Rule: There are several types of rules for performing di erent types of
evaluation of the supplied data. Currently we have minim:SoftwareEnvRule,
which tests to see if a particular piece of software is available in the
current execution environment, and minim:QueryTestRule, which uses a
querybased approach to assess the tness of a target.</p>
        <sec id="sec-3-2-1">
          <title>4 http://tools.ietf.org/html/rfc2119</title>
        </sec>
        <sec id="sec-3-2-2">
          <title>5 The namespace of minim is purl.org/minim/minim#.</title>
          <p>The following script, expressed using Turtle format, de nes an example Minim
checklist, which is to be used to assess each chemical compound must have
exactly one InChI number. The checklist has one requirement that must be satis ed
(line 9), i.e.,:InChI. The test of this rule is expressed by a SPARQL query (lines
19-20), which searches for the InChI identi er of a compound. The cardinality
rule (lines 22-23) speci es that there must be exactly 1 matching query result
associated with an evaluated compound.</p>
          <p>In the current checklist implementation the minim:QueryTestRule is used to
handle most of the checklist requirements we encounter. It can be associated with
two elements: a query pattern (minim:Query) (lines 16-26), which is evaluated
against the RDF data from the RO, and an optional external resource, which
contains additional RDF statements that may be needed to complete the
assessment. Every minim:QueryTestRule incorporates a minim:QueryResultTest,
which takes the query result (which in our current case, a SPARQL query
result) and returns a True (pass) or False (fail) result according to the type of test
performed. Currently our Minim model de nes 5 types of tests.</p>
          <p>{ minim:CardinalityTest, evaluates the minimum and/or maximum number
of distinct matches in the query result against the declared conditions.
{ minim:AccessibilityTest, evaluates whether a target resource indicated
by the query result is accessible, by for example performing an HTTP HEAD
request to the resource URI.
{ minim:AggregationTest, tests the presence of resources in an RO that is
used as the input to our assessment.
{ minim:RuleTest, de nes the additional rules to be applied to the assessment
results returned from the evaluation of another minim:QueryTestRule. In
this way, we can avoid writing too big rules and combine di erent types of
rules, for example a query test rule with a liveness test rule.
{ minim:ExistsTest, which can be used as a shortcut for a structure that
combines a minim:RuleTest and minim:CardinalityTest to evaluate the
existence of a particular resource in the evaluated data.</p>
          <p>The Minim model is a refactor of our previous work [14], which addressed
quality needs for enhancing scholarly communication (such as scenario 2). It
has been extended by 1) explicitly de ning an expandable set of test types;
and 2) providing extension points allowing de nitions of new assessment rules,
assessment tests, and types of queries used to perform query-based tests (see
Rule, Query and QueryResultTest in Figure 1).</p>
          <p>Clearly, not every measure of quality can be evaluated automatically. For
example, establishing correctness of stated facts may require independent
validation [13]. Our approach allows direct tests to be combined with such
independent validation or review, the latter of which may be simply expressed as
quality metadata about the target dataset. A systematic assessment of how our
checklist-based approach can support most of the existing known quality
dimensions is a key part of our future work. Our focus on extensibility allows new
automatic assessments to be introduced in a principled fashion. Examples of
checklists that combine automatic evaluation with manual review may be found
in our GitHub repository 6.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4 Implementation: The Minim Checklist Framework</title>
      <p>The checklist framework is implemented in Python as both a command-line tool,
ro-manager, and a RESTful service78. Source code is in GitHub9.</p>
      <p>As shown in Figure 2, the evaluation framework takes four inputs: a Research
Object (RO) that containing a set of RDF annotations, a Minim le, a purpose
indication, and an optional target resource URI (if not speci ed, the RO itself
is the target). The framework uses a checklist from the Minim le selected by
the purpose and target, applying each of the assessment tasks described by each
checklist item to the RDF graph presented by the Research Object.</p>
      <p>We chose SPARQL to express the QueryTestRules within a Minim checklist,
as SPARQL is a widely available standard for querying and accessing RDF data.
Our comparison with other semantic web technology choices is presented in
Section 6.</p>
      <p>The assessment result contains quite extensive content in the form of an RDF
graph. For web applications using these results, our implementation provides two
additional services that return JSON or HTML checklist results that facilitate
presentation of a more user-friendly \tra c-light"display, with \green ticks" for
satis ed requirements, and \red crosses" and \yellow crosses" meaning failure
of a MUST and SHOULD requirement respectively.</p>
      <sec id="sec-4-1">
        <title>6 https://github.com/wf4ever/ro-catalogue/tree/master/minim</title>
      </sec>
      <sec id="sec-4-2">
        <title>7 http://purl.org/minim/checklist-service</title>
      </sec>
      <sec id="sec-4-3">
        <title>8 Example REST service use is at https://github.com/wf4ever/ro-catalogue/</title>
        <p>blob/master/minim/REST-invoke-checklist.sh</p>
      </sec>
      <sec id="sec-4-4">
        <title>9 https://github.com/wf4ever/ro-manager/</title>
        <p>In this section we show how the two motivating scenarios can be supported by
our checklist tool. All the resources used for these case studies can be accessed
in our Github repository10. Our exercise shows that our model and tool can
su ciently support assessment tasks from diverse domains, and at the same
time, enable an explicit representation of the quality requirements from these
tasks, which themselves can be valuable asset to a community.</p>
        <sec id="sec-4-4-1">
          <title>Assess quality of scienti c data using community checklist</title>
          <p>In the rst practical assessment we show how our checklist tool can be used to
express existing community checklists from scienti c domains in order to identify
any potential quality issues of a scienti c linked dataset. This actually reproduces
the assessment by the previous MIM study [7] in our rst motivating scenario.
We reuse the chemical compound linked data and the checklist requirements
de ned in that study.</p>
          <p>In that study 11 quality requirements were de ned, based on a guideline from
the chemistry domain. We analysed the tests required by each requirement11
and categorised them into 3 di erent types: existence of information, type of
information present, and cardinality of values provided. Our Minim model can
be used to express these types of test, and the complete Mimim representation
of these requirements is in our Github repository. We applied this checklist to
100 (limited by a performance constraint of the RO access mechanism used,
currently being addressed) of the total 7,572 chemical compounds used in [7]
and our checklist tool was able to reproduce exactly the same assessment result
10 http://purl.org/minim/in-use-submission/
11 https://github.com/wf4ever/ro-catalogue/blob/master/v0.1/
minim-evaluation/checklist-item-survey.md
as the MIM checklist tool. Whilst we see this limited assessment as su cient
to demonstrate that we can reproduce the results of the MIM checklist, future
work (discussed in Section 8) will include a full validation for completeness.</p>
        </sec>
        <sec id="sec-4-4-2">
          <title>Assess quality of scholarly communication research objects for speci c purpose</title>
          <p>In our second case study we apply our checklist tool to a set of scienti c work ows
from the myExperiment.org repository. These work ows commonly rely on a
third-party bioinformatics Web service provided by a research organisation in
Japan12. At the end of year 2012, they announced that these services which
were available as WSDL service would be upgraded to RESTful services and
the WSDL service endpoints would no longer be supported, leading to failure of
dependent work ows. Although it is impossible for them to be executable after
the service upgrade, our assessment can enhance the quality of documentations
about these work ows so that they can at least be understandable, repairable,
and veri able in the future.</p>
          <p>Therefore, we designed a speci c checklist, based on our previous analysis
of causes to work ow quality issues [14]. In the checklist we de ne a list of
requirements to be assessed, including: the presence of all input data; the presence
of the work ow de nition le; the presence of provenance logs of previous runs;
and the accessibility of all the Web services used in a work ow.</p>
          <p>22 work ows from myExperiment.org were applicable to our test. Our
assessment managed to ensure that all the required information was associated
with each work ow (see the full assessment result in our Github repository).
After the service update took place, our checklist tool was able to successfully
detect quality degradation for all the work ows and highlight explicitly the set
of problematic services which caused the work ow no longer executable (see an
example assessment result13). The assessment can be reproduced using resources
in our Github repository.
6</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Discussions</title>
      <p>As an approach that is substantially based on semantic web technologies, the
goals and features of our checklist-based framework can be seen to overlap
with some major semantic web technologies like the Web Ontology Language
(OWL) 14 and SPIN15, which have been considered in our design process.
However, our focus was to provide a higher level data model, which can more directly
re ect quality requirements from users or speci c scenarios. Although these
semantic web technologies can be complementary to our approach, they cannot in
isolation (fully) support all the quality assessment requirements identi ed from
our scenarios.
12 http://www.genome.jp/kegg/
13 http://tinyurl.com/btxdlmv - this is a live service link
14 http://www.w3.org/TR/owl2-overview/
15 http://spinrdf.org
6.1</p>
      <sec id="sec-5-1">
        <title>Comparison with an OWL-based Approach</title>
        <p>OWL ontologies support the description of classes that detail the features
necessary for an individual data item to be a member of that class. These class
descriptions are analogous to the description of requirements in our checklist.
OWL also has an RDF serialisation and extends RDF semantics16 to operate
over RDF data. We can express our InChI requirement in OWL as follows:
1 Class : InChI
2 SubClassOf : chembox : StdInChI some : InChIValue .</p>
        <p>However, the current OWL 2 RDF semantics contain two features that are
incompatible with our quality checking scenario:
{ The Open World Assumption (OWA). If an InChI were to be de ned without
a corresponding InChIValue, this would not be highlighted as an error by an
OWL reasoner. Instead the OWA results in the inference that there exists
an InChIValue, but that it is currently unknown. This directly con icts with
our need for an existence check.
{ No Unique Names Assumption. We can extend the above requirement to
include a cardinality restriction to say that there must be one and only one
InChIValue. The presence of two di erent InChI values would not however
raise an error. Instead the assumption would be made that the two
InChIValues are in fact the same. This directly con icts with our need for cardinality
checks in a quality assessment scenario.</p>
        <p>An alternative to the traditional OWL 2 Semantics are Integrity Constraint
Semantics (ICs)17. ICs are a semantics for OWL that employ a Closed World
Assumption as well as a form of the Unique Names assumption. These semantics
therefore allow the use of OWL classes to be interpreted as integrity constraints.
The Stardog database18 currently provides an implementation of OWL with ICs.</p>
        <p>
          One practical implementation of ICs is achieved by transforming the OWL
classes to SPARQL queries. Each axiom in an OWL IC Ontology is transformed
into a corresponding SPARQL query. This ability to realise ICs as SPARQL
queries implies that by supporting a SPARQL based approach for requirement
description, Minim achieves at least some of the expressiveness as an approach
based upon OWL ICs. However, a purely OWL ICs based approach presents a
number of restrictions with respect to what can be expressed in our requirements:
{ Expression of di erent requirement levels such as MUST, SHOULD, and
MAY. OWL IC semantics are primarily concerned with binary satis ability,
where we capture more nuanced levels of satisfaction. We believe would be
more di cult to create checklists in OWL that capture these.
16 http://www.w3.org/TR/rdf-mt/#MonSemExt
17 http://stardog.com/docs/sdp/icv-specification.html
18 http://www.stardog.com/
{ Flexibility and extensibility to perform broader resource accessibility and
software environment tests that can be supported by our Minim tool. For
example verifying the web-accessibility of work ow input les lies outside
the expressive scope of OWL (though might conceivably be handled through
the introduction of new primitive classes and OWL resoner extensions).
{ Expressing rules that validate data literal values. This has previously been
highlighted as a restriction of an OWL based approach to data validation in
the life sciences [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ].
6.2
        </p>
      </sec>
      <sec id="sec-5-2">
        <title>Comparison with a SPIN-based Approach</title>
        <p>SPIN iprovides a query-based modelling language to express rules and logical
constraints over RDF data. It is used by the previously discussed MIM
checklistbased assessment framework.</p>
        <p>The property of spin:constraint can support a set of features in common
with our Minim tool. spin:constraint can be associated with an rdfs:Class,
e.g. chembox:InCHI, and de nes the constraints that instances of the class should
comply with. The constraints can be expressed using SPARQL ASK or
CONSTRUCT queries that are expressed using SPIN syntax in RDF. This structure
can be used to support most of our query-based tests, apart from the accessibility
tests. Additionally, spin:Template, which provides a meta-modelling function
to group SPARQL queries so that they can be reused, is very similar to the role
of minim:Rule in our model. However, at the time of writing, SPIN was not yet
established as a standard and implementations of SPIN engines were limited. A
purely SPIN-based approach also shares the rst two restrictions as an OWL
ICs based approach, as analysed above.
6.3</p>
      </sec>
      <sec id="sec-5-3">
        <title>Summary</title>
        <p>OWL, OWL ICs, and SPIN are clearly complementary to our Minim model
approach. Although they cannot be directly used to support expressing quality
assessment requirements, they can complement our SPARQL-based
implementation of the checklist tool. SPARQL was chosen for our tool implementation
because it is a more established standard for querying RDF data, with a
number of known implementations. Combined with our Minim model, SPARQL can
support all the expression of constraints and most of the inference functions as
SPIN. However, our Minim model can also be extended and implemented using
these alternative technologies. The minim:Query class is one extensition point
for supporting SPIN-like queries, and minim:Rule can be extended to de ne
other than query-based test rules.
7</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Related Work</title>
      <p>
        Zaveri et al. [13] provides a timely and extensive survey on quality assessment of
linked data. The survey is mainly organised by quality dimensions rather than
the actual methodologies used by the reviewed works. Of the 21 works included
in the review, a larger portion of them are based on speci c algorithms, such as
the trust evaluation by Golbeck [8] , or use a dimension-driven approach, such as
Bizer et al [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], or take a purpose-built approach to provide solutions to a speci c
problem in a speci c application scenario, such as Gueret et al. [9]. 3 of the works
take an approach more closely related to ours by supporting an explicit
expression of quality requirements. However, the quality schema provided by Sieve [12]
is rather simple, mainly targeted to express the con guration parameters and
the functions to be used for the assessment; and the quality ontologies proposed
by SemRef [11] and SWIQA [6] are based on a series of quality dimensions.
8
      </p>
    </sec>
    <sec id="sec-7">
      <title>Conclusions and Future Work</title>
      <p>Quality assessment is a paramount issue in supporting the successful re-use of
Scienti c Linked Data. Not being able to express speci c quality assessment
requirements according to the needs from speci c assessment tasks has been a
bottleneck to the quality enhancement of linked data resources. To ll in this
critical gap, we propose a checklist-based approach that allows explicit expression
of quality requirements that can directly re ect users' needs from their concrete
quality assessment tasks, and at the same provides exible extensibility to cope
with new needs. We show how our approach can support two exemplar case
studies from scienti c domains. We learnt valuable lessons about how various
state-of-the-art semantic web technologies could support our concrete use in
practice. The very lightweight SPARQL-based implementation has shown great
promise in supporting these practical needs.</p>
      <p>Our next steps will focus on the extensibility of the tool architecture, by
exploring the possibility of a plug-in framework to enable plugging-in of third-party
services. We are also prototyping a user interface tool to facilitate the creation
of Minim checklists. Finally we are planning a systematic mapping between the
existing quality dimensions and the constructs available in our checklist data
model, to extend the function evaluation of our model.
5. D. De Roure, C. Goble, and R. Stevens. The design and realisation of the
myexperiment virtual research environment for social sharing of work ows. Future
Generation Computer Systems, 25:561{567, 2009.
6. Christian Furber and Martin Hepp. Swiqa{a semantic web information quality
assessment framework. In Proceedings of European Conference on Information
Systems, 2011.
7. Matthew Gamble, Carole Goble, Graham Klyne, and Jun Zhao. Mim: A minimum
information model vocabulary and framework for scienti c linked data. In
EScience (e-Science), 2012 IEEE 8th International Conference on, pages 1{8. IEEE,
2012.
8. Jennifer Golbeck and Aaron Mannes. Using trust and provenance for content
ltering on the semantic web. In Proceedings of the Models of Trust for the Web
Workshop, 2006.
9. Christophe Gueret, Paul Groth, Claus Stadler, and Jens Lehmann. Assessing linked
data mappings using network measures. In Proceedings of the European Semantic
Web Conference, pages 87{102. Springer, 2012.
10. Brigette M. Hales and Peter J. Pronovost. The checklist{a tool for error
management and performance improvement. Journal of critical care, 21(3):231{235,
2006.
11. Yuangui Lei, Victoria Uren, and Enrico Motta. A framework for evaluating
semantic metadata. In Proceedings of the 4th international conference on Knowledge
capture, pages 135{142. ACM, 2007.
12. Pablo N Mendes, Hannes Muhleisen, and Christian Bizer. Sieve: linked data quality
assessment and fusion. In Proceedings of the 2012 Joint EDBT/ICDT Workshops,
pages 116{123. ACM, 2012.
13. Amrapali Zaveri, Anisa Rula, Andrea Maurino, Ricardo Pietrobon, Jens Lehmann,
and Soren Auer. Quality assessment methodologies for linked open data. Semantic
Web Journal. submitted on 12/14/2012.
14. Jun Zhao, Jose Manuel Gomez-Perez, Khalid Belhajjame, and et al. Why work ows
break-understanding and combating decay in taverna work ows. In IEEE eScience,
pages 1{8, 2012.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <given-names>Khalid</given-names>
            <surname>Belhajjame</surname>
          </string-name>
          , Oscar Corcho, Daniel Garijo, Jun
          <string-name>
            <surname>Zhao</surname>
          </string-name>
          , and et al.
          <article-title>Work owcentric research objects: First class citizens in scholarly discourse</article-title>
          .
          <source>In Proceeding of SePublica2012</source>
          , pages
          <fpage>1</fpage>
          {
          <fpage>12</fpage>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <given-names>Christian</given-names>
            <surname>Bizer</surname>
          </string-name>
          and Richard Cyganiak.
          <article-title>Quality-driven information ltering using the wiqa policy framework</article-title>
          .
          <source>Web Semantics: Science, Services and Agents on the World Wide Web</source>
          ,
          <volume>7</volume>
          (
          <issue>1</issue>
          ):1{
          <fpage>10</fpage>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <given-names>Jerven</given-names>
            <surname>Bolleman</surname>
          </string-name>
          , Alain Gateau, Sebastien Gehant, and
          <string-name>
            <given-names>Nicole</given-names>
            <surname>Redaschi</surname>
          </string-name>
          .
          <article-title>Provenance and evidence in uniprotkb</article-title>
          .
          <source>In Proceedings of the 3rd International Workshop on Semantic Web Applications and Tools for the Life Sciences</source>
          , Berlin, Germany,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <given-names>Sangyoon</given-names>
            <surname>Chin</surname>
          </string-name>
          , Kyungrai Kim, and
          <string-name>
            <surname>Yea-Sang Kim</surname>
          </string-name>
          .
          <article-title>A process-based quality management information system</article-title>
          .
          <source>Automation in Construction</source>
          ,
          <volume>13</volume>
          (
          <issue>2</issue>
          ):
          <volume>241</volume>
          {
          <fpage>259</fpage>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>