<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Representing verifiable statistical index computations as linked data</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Jose Emilio Labra Gayo</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Hania Farham</string-name>
          <email>hania@webfoundation.org</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Juan Castro Ferna´ndez</string-name>
          <email>juan.castrog@weso.es</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jose Mar´ıa A´ lvarez Rodr´ıguez</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Dept. Computer Science Carlos III University</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>The Web Foundation</institution>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>WESO Research Group</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>In this paper we describe the development of the Web Index linked data portal that represents statistical index data and computations. The Web Index is a multi-dimensional measure of the World Wide Web's contribution to development and human rights globally. It covers 81 countries and incorporates indicators that assess several areas like universal access; freedom and openness; relevant content; and empowerment. In order to empower the Web Index transparency, one internal requirement was that every published data could be externally verified. The verification could be that it was just raw data obtained from an external source, in which case, the system must provide a link to the data source or that the value has been internally computed, in which case, the system provides links to those values. The resulting portal contains data that can be tracked to its sources so an external agent can validate the whole index computation process. We describe the different aspects on the development of the WebIndex data portal, which also offers new linked data visualization tools. Although in this paper we concentrate on the Web Index development, we consider that this approach can be generalized to other projects which involve the publication of externally verifiable statistical computations.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1 Introduction</title>
      <p>Statistical indexes are a widely accepted practice that have been applied to numerous
domains like economics and Bibliometrics (Impact factor), research and academic
performance (H-Index or Shanghai rankings), cloud computing (Global Cloud Index, by
CISCO), etc. We consider that those indexes could benefit from a Linked Data approach
where the rankings could be seen, tracked and verified by their users linking each rank
to the original values and observations from which it has been computed.</p>
      <p>As a motivating example, we will employ the Web Index project (http://
thewebindex.org), which created an index to measure the World Wide Web’s
contribution to development and human rights globally. Scores are given in the areas of access;
freedom and openness; relevant content; and empowerment. First released in 2012, the
2013 Index has been expanded and refined to include 20 new countries and features an
enhanced data set, particularly in the areas of gender, Open Data, privacy rights and
security.</p>
      <p>
        The 2012 version offered a data portal4 whose data was obtained by transforming
raw observations and precomputed values from Excel sheets to RDF. The technical
description of that process was described in [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] where we followed the methodology
presented in [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>In this paper, we describe the development of the 2013 version of that data portal,
where we employ a new validation and computation approach that enables the
publication of a verifiable linked data version of WebIndex results.</p>
      <p>
        We defined a generic vocabulary of computational index structures called Computex
which could be applied to compute and validate any other kind of index and can be seen
as an specialization of the RDF Data Cube vocabulary [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
      </p>
      <p>Given that the most important part of a data portal about statistical indexes are the
numeric values of each observation we established the internal requirement that any
value published should be justified either declaring from where it had been obtained or
linking it to the values of other observations from which it had been computed.</p>
      <p>
        The validation process employs a combination of SPARQL [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] queries and Shape
Expressions [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] to check the different integrity constraints and computation steps in
a declarative way. The resulting data portal http://data.webfoundation.org/webindex/
2013 contains not only a linked data view about the statistical data but also a machine
verifiable justification of the index ranks.
      </p>
      <p>In the rest of the paper we will use Turtle and SPARQL notation and assume that
the namespaces have been declared using the most common prefixes found in http:
//prefix.cc.
2</p>
    </sec>
    <sec id="sec-2">
      <title>WebIndex Computation Process</title>
      <p>The Web Index is a composite measure that summarizes in a single (average)
number the impact and value derived from the Web in various countries. There are serious
challenges when attempting to measure and quantify some of the dimensions the Index
covers (e.g. the social and political), and suitable proxies were used instead.</p>
      <p>Two types of data were used in the construction of the Index: existing data from
other data providers (secondary data), and new data gathered via a multi-country
questionnaire (primary data) which was specifically designed by the Web Foundation and
its advisers. These primary data will begin to fill in some of the gaps in measurement
of the utility and impact of the Web in various countries.</p>
      <p>As the Web Index covers a large number of countries, some of which have serious
data deficiencies or were not covered by the data providers, some missing data had to
be imputed.</p>
      <p>The following steps summarise the computation process of the Index:
1. Take the data for each indicator from the data source for the 81 countries covered
by the Index for the 2007-2012 time period (or 2013, in the case of the Web Index
expert assessment survey).
4 http://data.webfoundation.org
2. Impute missing data for every secondary indicator for the sample of 81 countries
over the period 2007-2012. Broadly, the imputation of missing data was done using
two methods: country-mean substitution if the missing number is in the middle
year (e.g. have 2008 and 2010 but not 2009), or taking arithmetic growth rates on a
year-by-year basis.
3. Normalise the full (imputed) dataset using z-scores, making sure that for all
indicators, a high value is good and a low value is bad.
4. Cluster some of the variables, taking the average of the clustered indicators
postnormalisation. For the clustered indicators, this clustered value is the one to be used
in the computation of the Index components.
5. Compute the component scores using arithmetic means, using the clustered values
where relevant.
6. Compute the min-max values for each z-score value of the components, as this is
what will be shown in the visualisation tool and other publications containing the
component values (generally, it is easier to understand a min-max number in the
range of 0 – 100 rather than a standard deviation-based number). The formula for
this is: mxaxmminin 100
7. Compute sub-index scores by calculating the weighted averages of the relevant
components for each sub-Index and the min-max values for each z-score value of
the sub-Indexes.
8. Compute overall composite scores by calculating the weighted average of the
subindexes and the min-max values.</p>
      <p>The computation process was originally done by human experts using an Excel file
although once the process was established, the computation was automated to validate
the whole process.
3</p>
    </sec>
    <sec id="sec-3">
      <title>WebIndex workflow</title>
      <p>The WebIndex workflow has been depicted in figure 1. The Excel file was comprised
of 184 Excel sheets and contained a combination of raw, imputed and normalized data
created by the statistical experts.</p>
      <p>That external data was filtered and converted to RDF by means of an specialized
web service called wiFetcher5.</p>
      <p>Although some of the imported values had been pre-computed in Excel by human
experts, we collected only the raw values, so we could automatically compute and
validate the results.</p>
      <p>In this way, another application called wiCompute6 took the raw values and
computed the index following the computation steps defined by the experts. wiCompute
carried out the computations generating RDF datasets for the intermediary results and
linking the generated values to the values from which they had been computed.</p>
      <p>Finally, the RDF data generated was published to a SPARQL endpoint from which
we created a specialized visualization tool called Wesby7.</p>
      <sec id="sec-3-1">
        <title>5 https://github.com/weso/wiFetcher 6 https://github.com/weso/wiCompute 7 https://github.com/weso/wesby</title>
        <p>Given the statistical nature of the data, the WebIndex data model is based on the RDF
Data Cube vocabulary. Figure 4 represents the main concepts of the data model</p>
        <p>As can be seen, the main concept are observations of type qb:Observation, which
can be raw observations, obtained from an external source, or computed observations
derived from other observations. Each observation has a float value cex:value and is
related to a country, a year, a dataset and an indicator.</p>
        <p>A dataset contains a number of slices, each of which also contains a number of
observations.</p>
        <p>
          Indicators are provided by an organization of type org:Organization from the
Organization ontology[
          <xref ref-type="bibr" rid="ref18">18</xref>
          ]. Datasets are also published by organizations.
        </p>
        <p>As a sample of some data, an observation can be that Italy has the normalized value
using Z-Scores of 0:80 in 2007 for the indicator WEF L (Impact of ICT on
organizational models) provided by the World Economic Forum. This information can be
represented in RDF using Turtle syntax as8:
obs:computed_26549 a qb:Observation ;
cex:indicator indicator:WEF_L ;
qb:dataSet dataset:d_52 ;
cex:value "-0.80"ˆˆxsd:double ;
cex:ref-area country:Italy ;
cex:ref-year 2007 ;
sdmx-concept:obsStatus cex:Normalized ;
cex:computation computation:c26550
...other properties omitted for brevity
8 The real observation is http://data.webfoundation.org/webindex/v2013/observation/computed
2007 1386752461095 26549. The real URIs also include an internal long number used to
uniquely identify each entity</p>
        <p>Notice that the WebIndex data model contains data that is completely interrelated.
Observations are linked to indicators, datasets and computations. Datasets contain also
links to slices and slices have links to indicators and observations again. Both datasets
and indicators are linked to the organizations that publish or provide them.</p>
        <p>The following example contains a sample of interrelated data for this domain.</p>
        <p>Computed observations and datasets contain a property cex:computation that
associate them to a node of type cex:Computation which links the computed
observation to the observations from which it has been obtained. In the above example, the
computation c26550 indicates that it is a normalization of the observation obs:obs29761
using the observations in slice slice:WEF_L2007-Imputed which has a standard
deviation of 0:75 and a mean of 4:39. Including these declarations, an external agent can
verify if the value of the observation has been well computed or if it has been
tampered. We also noticed that these declarations had another positive effect to debug the
computation process in the development phase of the data portal.
5</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Computex vocabulary</title>
      <p>The Computex9 defines terms related to the computation of statistical index data and is
compatible with RDF Data Cube vocabulary. Some terms defined in the vocabulary are:
– cex:Concept represents the entities that we are indexing. In the case of the Web
Index project, the concepts are the different countries. In other applications it could
be Universities, journals, services, etc.
– cex:Indicator. A dimension whose values add information to the Index.
Indicators can be simple dimensions, for example: the mobile phone suscriptions per 100
population, or can be composed from other indicators.
– cex:Computation. It represents a computation. We included the main
computation types that we needed for the WebIndex project, which have been summarized
in Table 5. That list of computation types is non-exhaustive and can be further
extended in the future.
– cex:WeightSchema a weight schema for a list of indicators. It consists of a weight
associated for each indicator which can be used to compute an aggregated
observation.
6</p>
    </sec>
    <sec id="sec-5">
      <title>Development and Validation approach</title>
      <p>The validation approach employed in the 2012 WebIndex project was based on ad-hoc
resource templates and a MD5 checksum field. Apart from that, we did not verify that</p>
      <sec id="sec-5-1">
        <title>9 http://purl.org/weso/ontology/computex</title>
        <p>Mean
Increment
Copy
Z-score
Ranking
AverageGrowth
WeightedMean
Properties
Description
No computation. Raw value obtained from
external source.</p>
        <p>Mean of a set of observations
cex:observation
cex:slice
Increment an observation by a given amount cex:observation
cex:amount
A copy of another observation cex:observation
A normalization of an observation using the val- cex:observation
ues from a Slice. cex:slice
Position in the ranking of a slice of observations. cex:observation
cex:slice
Expected average growth of N observations cex:observations10
Weighted mean of an observation cex:observation
cex:slice
cex:weightSchema
the precomputed values imported from the Excel sheets really matched the value that
could be obtained by following the declared computation process.</p>
        <p>In the 2013 version, we did a step forward on the validation approach. The goal was
not only to check that a resource contained a given set of fields and values, but also that
those values really matched the values that can be obtained by following the declared
computations.</p>
        <p>The proposed approach was inspired by the integrity constraint specification
proposed by the RDF Data Cube vocabulary, which employs a set of SPARQL ASK queries
to check the integrity of RDF Data Cube data. Although ASK queries provide a good
means to check integrity, in practice their boolean nature does not offer too much help
when a dataset does not accomplish with the data model.</p>
        <p>We decided to use CONSTRUCT queries which, in case of error, contain an error
message and a list of error parameters that can help to spot the problematic data.</p>
        <p>
          We transformed the ASK queries defined in the RDF Data Cube specification to
CONSTRUCT queries. In order to make our error messages compatible with EARL [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ],
we have defined cex:Error as a subclass of earl:TestResult and declared it to
have the value earl:failed for the property earl:outcome.
        </p>
        <p>We have also created our own set of SPARQL CONSTRUCT queries to validate the
Computex vocabulary terms, specially the computation of index data. For example, the
following query validates whether every observation has at most one value.
CONSTRUCT { [ a cex:Error ; cex:errorParam # ... omitted
cex:msg "Observation has two different values" . ]
} WHERE { ?obs a qb:Observation .
?obs cex:value ?value1 . ?obs cex:value ?value2 .</p>
        <p>FILTER ( ?value1 != ?value2 ) }</p>
        <p>Using this approach, it is possible to define more expressive validations. For
example, we are able to validate whether an observation has been obtained as the mean of
other observations.</p>
        <p>CONSTRUCT { [ a cex:Error ; cex:errorParam # ...omitted
cex:msg "Mean value does not match" ] .
} WHERE { ?obs a qb:Observation ;
cex:computation ?comp ;
cex:value ?val .</p>
        <p>?comp a cex:Mean .
{ SELECT (AVG(?value) as ?mean) ?comp WHERE {
?comp cex:observation ?obs1 .</p>
        <p>?obs1 cex:value ?value ;
} GROUP BY ?comp }
FILTER( abs(?mean - ?val) &gt; 0.0001) }</p>
        <p>
          Validating statistical computations using SPARQL queries offered a good exercise
to check SPARQL expressiveness. Although we were able to express most of the
computation types, some of them had to employ functions that were not part of SPARQL
1.1 or had to be defined in a limited way. We described these limits in [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ].
        </p>
        <p>We implemented an online validation tool called Computex11 which takes as input
an RDF graph and checks if it follows the integrity constraints defined by
Computex. The validation tool can also check if the RDF graph follows the RDF Data Cube
integrity constraints and it can also do the index computation for RDF Graphs.
Although this declarative approach was very elegant, computing the webindex using only
SPARQL queries was not practical (it took around 15 minutes for a small subset), so
the computation process was finally done by a specialized program implemented in
Scala 12.
7</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Visualizing the data portal</title>
      <p>
        We developed a visualization tool called Wesby 13 which takes as input an SPARQL
endpoint and offers a linked data browsing experience. Wesby was inspired by
Pubby [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] and was developed in Scala using the Play! Framework. Wesby combines
the visualization with a set of templates to offer specialized views for different types of
resources. For example, figure 3 contains the WebIndex visualization of Italy14. The
interactive visualization graphics use a javascript library called WesCountry that we have
also developed 15.
      </p>
      <p>When there is no template for a given type of node, Wesby shows a table of
properties and values similar to Pubby. Wesby also handles content negotiation so it can return
different representations depending on the ACCEPT header.
11 http://computex.herokuapp.com/
12 Source code is available here: https://github.com/weso/wiCompute
13 http://wesby.weso.es
14 It can be seen here: http://data.webfoundation.org/webindex/v2013/country/ITA
15 http://weso.github.io/wesCountry/</p>
      <p>In order to document the resulting data portal we created a set of templates using
Shape Expressions 16. We consider that this approach offers a good balance between
human readability and machine processable specification.
8</p>
    </sec>
    <sec id="sec-7">
      <title>Related work</title>
      <p>
        There is a growing interest in developing solutions to improve the quality of linked
data [
        <xref ref-type="bibr" rid="ref11 ref12 ref15">11, 15, 12</xref>
        ]. We consider that it is very important to publish linked data that is
not only of high quality, but also that can automatically be validated. Validating RDF
has also attracted a number of approaches. Most of them were presented at the W3c
Workshop on RDF Validation [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] and can be classified as inference based, SPARQL
queries or grammar based.
      </p>
      <p>
        Inference based approaches try to adapt OWL for validation proposes. However,
the use of Open World and Non-unique name assumption limits the validation
possibilities. A variation of OWL semantics using Closed World Assumption to express
integrity constraints has been proposed in [
        <xref ref-type="bibr" rid="ref16 ref20 ref6">6, 20, 16</xref>
        ]. SPARQL queries can also express
validation constraints and offer a great level of expressiveness [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. Grammar based
approaches like OSLC Resource Shapes [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] and Dublic Core Application Profiles [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]
define a domain specific language to declare the validation rules. Recently, Shape
Expressions [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] have been proposed as a new technology to describe and validate RDF
data portals.
      </p>
      <p>
        Representing statistical linked data has also seen an increasing interest. SDMX 17 is
the primary format of the main statistical data organizations. The transformation of
16 http://weso.github.io/wiDoc/
17 http://sdmx.org/
SDMX-ML to RDF/XML has been described in [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. The RDF Data Cube
vocabulary [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] has been accepted as a W3c Recommendation technology to publish
multidimensional statistical data and to link it with other concepts and data. We have opted
to follow the RDF Data Cube vocabulary and in fact, we consider that Computex can
be seen as a further specialization of RDF Data Cube to represent statistical index
computations.
      </p>
      <p>
        Another line of related work is the representation of mathematical expressions as
linked data. Lange [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] gives an overview of the different approaches. OpenMath was
proposed as an extensible standard that can represent the semantic meaning of
mathematical objects. Wenzel and Reinhardt [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ] propose an approach to integrate OpenMath
with RDF data for the representation of mathematical relationships and the integration
of mathematical computations into reasoning systems. We consider Computex as a first
step in that direction to represent statistical computations and we expect more future
work to appear about how to represent statistical computations as linked data.
9
      </p>
    </sec>
    <sec id="sec-8">
      <title>Conclusions</title>
      <p>In this paper, we described how we were able to represent statistical index computations
as linked data which include information to track the origin of any published
observation. Although the number of triples were around 3,5 million, we consider that the data
portal is of medium size, so we were able to play with different validation possibilities.</p>
      <p>Although we have been able to express most of the computations using SPARQL
queries, we have found some limitations in current SPARQL 1.1 expressiveness with
regards to built-in functions on maths, strings, RDF Collections and performance. In
fact, although we initially wanted to do the whole computation process using SPARQL
CONSTRUCT queries, we found that it took longer than expected and was difficult to
debug, so we opted to develop an independent program that did all the computation
process in a few seconds.</p>
      <p>After participating in the W3c RDF Validation workshop we were attracted by the
Shape Expressions formalism so we developed the documentation of the WebIndex data
portal using Shape Expressions. We consider that some structural parts of the data portal
can be better expressed in Shape Expressions.</p>
      <p>Our future work is to automate the declarative computation of index data from the
raw observations and to check the performance using the Web Index data. We are also
improving the Wesby visualization tool and the WesCountry library for statistical
graphics. We are even considering to relate visualization templates with Shape Expressions
offering a better separation of concerns in the development process.
10</p>
    </sec>
    <sec id="sec-9">
      <title>Acknowledgements</title>
      <p>We would like to thank Jules Clements, Karin Alexander, Ce´sar Luis Alvargonza´lez,
Ignacio Fuertes Bernardo and Alejandro Montes for their collaboration in the
development of the WebIndex project.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S.</given-names>
            <surname>Abou-Zahra</surname>
          </string-name>
          .
          <source>Evaluation and Report Language EARL 1</source>
          .
          <article-title>0 schema</article-title>
          . http://www.w3.org/TR/EARL10-Schema/,
          <year>2011</year>
          . W3C Working Draft.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Alvarez</surname>
          </string-name>
          <article-title>Rodr´ıguez</article-title>
          , J. Clement,
          <string-name>
            <given-names>J. E. Labra</given-names>
            <surname>Gayo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Farhan</surname>
          </string-name>
          , and
          <string-name>
            <given-names>P.</given-names>
            <surname>Ordon</surname>
          </string-name>
          <article-title>˜ez. Cases on Open-Linked Data and Semantic Web Applications, chapter Publishing Statistical Data following the Linked Open Data Principles: The Web Index Project</article-title>
          ., pages
          <fpage>199</fpage>
          -
          <lpage>226</lpage>
          . IGI Global,
          <year>2013</year>
          . doi:
          <volume>10</volume>
          .4018/978-1-
          <fpage>4666</fpage>
          -2827- 4.
          <year>ch011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>I.</given-names>
            <surname>Boneva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. E.</given-names>
            <surname>Labra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hym</surname>
          </string-name>
          ,
          <string-name>
            <surname>E. G.</surname>
          </string-name>
          <article-title>Prud'hommeau, H</article-title>
          . Solbrig, and
          <string-name>
            <given-names>S.</given-names>
            <surname>Staworko</surname>
          </string-name>
          .
          <article-title>Validating RDF with Shape Expressions</article-title>
          . ArXiv e-prints,
          <source>Apr</source>
          .
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S.</given-names>
            <surname>Capadisli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Auer</surname>
          </string-name>
          , and A.
          <string-name>
            <surname>-C. Ngonga Ngomo</surname>
          </string-name>
          .
          <article-title>Linked sdmx data</article-title>
          .
          <source>Semantic Web Journal</source>
          , pages
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>F. A.</given-names>
            <surname>Cifuentes Silva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Sifaqui</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J. E. Labra</given-names>
            <surname>Gayo</surname>
          </string-name>
          .
          <article-title>Towards an architecture and adoption process for linked data technologies in open government contexts: a case study for the library of congress of chile</article-title>
          . In C. Ghidini,
          <string-name>
            <given-names>A.-C. N.</given-names>
            <surname>Ngomo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. N.</given-names>
            <surname>Lindstaedt</surname>
          </string-name>
          , and T. Pellegrini, editors,
          <string-name>
            <surname>I-SEMANTICS</surname>
          </string-name>
          , ACM International Conference Proceeding Series, pages
          <fpage>79</fpage>
          -
          <lpage>86</lpage>
          . ACM,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>K.</given-names>
            <surname>Clark</surname>
          </string-name>
          and
          <string-name>
            <given-names>E.</given-names>
            <surname>Sirin</surname>
          </string-name>
          .
          <article-title>On RDF validation, stardog ICV, and assorted remarks</article-title>
          .
          <source>In RDF Validation Workshop. Practical Assurances for Quality RDF Data</source>
          , Cambridge, Ma, Boston,
          <year>September 2013</year>
          . W3c, http://www.w3.org/
          <year>2012</year>
          /12/rdf-val.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>K.</given-names>
            <surname>Coyle</surname>
          </string-name>
          and
          <string-name>
            <given-names>T.</given-names>
            <surname>Baker</surname>
          </string-name>
          .
          <article-title>Dublin core application profiles. separating validation from semantics</article-title>
          .
          <source>In RDF Validation Workshop. Practical Assurances for Quality RDF Data</source>
          , Cambridge, Ma, Boston,
          <year>September 2013</year>
          . W3c, http://www.w3.org/
          <year>2012</year>
          /12/rdf-val.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>R.</given-names>
            <surname>Cyganiak</surname>
          </string-name>
          and
          <string-name>
            <given-names>C.</given-names>
            <surname>Bizer</surname>
          </string-name>
          .
          <article-title>Pubby: A linked data frontend for sparql endpoints</article-title>
          . http://www4.wiwiss.fu-berlin.de/pubby/.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>R.</given-names>
            <surname>Cyganiak</surname>
          </string-name>
          and
          <string-name>
            <given-names>D.</given-names>
            <surname>Reynolds</surname>
          </string-name>
          .
          <article-title>The RDF Data Cube Vocabulary</article-title>
          . http://www.w3.org/TR/vocab-data-cube/,
          <year>2013</year>
          . W3c Candidate Recommendation.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>S.</given-names>
            <surname>Harris</surname>
          </string-name>
          and
          <string-name>
            <given-names>A.</given-names>
            <surname>Seaborne</surname>
          </string-name>
          .
          <source>SPARQL 1</source>
          .
          <article-title>1 Query Language</article-title>
          . http://www.w3.org/TR/sparql11-query/,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>A.</given-names>
            <surname>Hogan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Harth</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Passant</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Decker</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Polleres</surname>
          </string-name>
          .
          <article-title>Weaving the pedantic web</article-title>
          .
          <source>In Linked Data on the Web Workshop (LDOW2010)</source>
          at WWW'
          <year>2010</year>
          , volume
          <volume>628</volume>
          , pages
          <fpage>30</fpage>
          -
          <lpage>34</lpage>
          . CEUR Workshop Proceedings,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>D.</given-names>
            <surname>Kontokostas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Westphal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Auer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hellmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Lehmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Cornelissen</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Zaveri</surname>
          </string-name>
          .
          <article-title>Test-driven evaluation of linked data quality</article-title>
          .
          <source>In Proceedings of the 23rd International Conference on World Wide Web, WWW '14</source>
          , pages
          <fpage>747</fpage>
          -
          <lpage>758</lpage>
          , Republic and Canton of Geneva, Switzerland,
          <year>2014</year>
          . International World Wide Web Conferences Steering Committee.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>J. E.</given-names>
            <surname>Labra</surname>
          </string-name>
          and
          <string-name>
            <surname>J. M. Alvarez</surname>
          </string-name>
          <article-title>Rodr´ıguez. Validating statistical index data represented in RDF using SPARQL queries</article-title>
          .
          <source>In RDF Validation Workshop. Practical Assurances for Quality RDF Data</source>
          , Cambridge, Ma, Boston,
          <year>September 2013</year>
          . W3c, http://www.w3.org/
          <year>2012</year>
          /12/rdf-val.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>C.</given-names>
            <surname>Lange</surname>
          </string-name>
          .
          <article-title>Ontologies and languages for representing mathematical knowledge on the semantic web</article-title>
          .
          <source>Semantic Web</source>
          ,
          <volume>4</volume>
          (
          <issue>2</issue>
          ):
          <fpage>119</fpage>
          -
          <lpage>158</lpage>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>P. N.</given-names>
            <surname>Mendes</surname>
          </string-name>
          , H. Mu¨hleisen, and
          <string-name>
            <given-names>C.</given-names>
            <surname>Bizer</surname>
          </string-name>
          . Sieve:
          <article-title>Linked data quality assessment and fusion</article-title>
          .
          <source>In Proceedings of the 2012 Joint EDBT/ICDT Workshops, EDBTICDT '12</source>
          , pages
          <fpage>116</fpage>
          -
          <lpage>123</lpage>
          , New York, NY, USA,
          <year>2012</year>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>B.</given-names>
            <surname>Motik</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Horrocks</surname>
          </string-name>
          , and
          <string-name>
            <given-names>U.</given-names>
            <surname>Sattler</surname>
          </string-name>
          .
          <article-title>Adding Integrity Constraints to OWL</article-title>
          . In C. Golbreich,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kalyanpur</surname>
          </string-name>
          , and B. Parsia, editors,
          <source>OWL: Experiences and Directions</source>
          <year>2007</year>
          (
          <article-title>OWLED 2007)</article-title>
          , Innsbruck, Austria, June 6-7
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>RDF</given-names>
            <surname>Working Group</surname>
          </string-name>
          <article-title>W3c</article-title>
          .
          <article-title>W3c validation workshop. practical assurances for quality rdf data</article-title>
          ,
          <year>September 2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>D.</given-names>
            <surname>Reynolds</surname>
          </string-name>
          . The Organization Ontology. http://www.w3.org/TR/vocab-org/,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>A. G.</given-names>
            <surname>Ryman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. L.</given-names>
            <surname>Hors</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Speicher</surname>
          </string-name>
          .
          <article-title>OSLC resource shape: A language for defining constraints on linked data</article-title>
          . In C. Bizer,
          <string-name>
            <given-names>T.</given-names>
            <surname>Heath</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Berners-Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hausenblas</surname>
          </string-name>
          , and S. Auer, editors,
          <source>Linked data on the Web</source>
          , volume
          <volume>996</volume>
          <source>of CEUR Workshop Proceedings. CEUR-WS.org</source>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>J.</given-names>
            <surname>Tao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Sirin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bao</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D. L.</given-names>
            <surname>McGuinness</surname>
          </string-name>
          .
          <article-title>Integrity constraints in OWL</article-title>
          .
          <source>In Proceedings of the 24th AAAI Conference on Artificial Intelligence (AAAI-10)</source>
          . AAAI,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>K.</given-names>
            <surname>Wenzel</surname>
          </string-name>
          and H. Reinhardt.
          <article-title>Mathematical computations for linked data applications with openmath</article-title>
          .
          <source>In Conferences on Intelligent Computer Mathematics, CICM</source>
          <year>2012</year>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>