<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Wikidata as a Challenge for Rule Systems</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Peter F. Patel-Schneider</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ege Atacan Doğan</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>New Jersey</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>U. S. A.</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Istanbul</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Turkey</string-name>
        </contrib>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <fpage>22</fpage>
      <lpage>24</lpage>
      <abstract>
        <p>Wikidata is a widely used knowledge graph with features that make it desirable for use in challenges for rule systems and subsequent benchmarks. The sheer size of Wikidata is itself a challenge for rule systems. Wikidata's large and broad ontology provides multiple areas for rule-based inference. Wikidata's continual updates provide a challenge for rule systems that retain consequences. Evaluating rule systems against the current data of Wikidata, rather than against a frozen version, prevents lock-in to a particular set of data and diminishing incremental advances against that data. Rule systems that perform well on Wikidata have the possibility of being incorporated into the software supporting Wikidata.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Current challenges and benchmarks for rule systems are generally based on synthetic data or frozen
extracts from existing data sources. We would like to see more evaluation of rule systems on actual,
current, independently produced data sources, instead of synthetic data or frozen extracts of data
sources. We feel that this kind of evaluation has several advantages that outweigh the disadvantages.</p>
      <p>Using actual, current, independently produced data sources eliminates the bias inherent in synthetic
data and the extraction process, in both size and data characteristics. It prevents lock-in to the frozen
synthetic generation process or the frozen data. It evaluates rule systems on data that they can actually
be run on in applications.</p>
      <p>We acknowledge that running on actual, current, independently produced data sources adds extra
complexity to evaluations. The first evaluation with a data source must provide a process that takes the
current version of the data and, if necessary, transform the data into a form that is suitable for evaluation.
This process must be able to accept not only the data current at the time of the first evaluation but
should also work for later versions of the data. All evaluations much provide a repeatable process
for running the system being evaluated not only on the processed data from when the system is first
being evaluated but also one that should work on later versions of the data. (Given the dificulties of
creating processes that work in the future, this efectively means that evaluation teams should commit
to providing help for future evaluations.)</p>
      <p>We further propose that an excellent data source for rule systems is Wikidata—not portions of
Wikidata but all of Wikidata. Wikidata is a large, complex knowledge graph with a large, complex
ontology, but not too large to be impossible to process for current rule systems. Wikidata is being
actively edited and is growing.</p>
      <p>Wikidata lacks a formal basis and also lacks rules, which might lead one to think that rule-based
systems have no place in Wikidata. But the intended meaning of many Wikidata constructs can be
formalized as rules in a suitable logic and Wikidata has a large number of constraints that can also be
formalized as rules. We feel that this implicit Wikidata suitable for the evaluation of rule-based systems
and can form the basis of challenges for rule-based systems.</p>
      <p>Wikidata-based evaluations can range from simply answering queries against Wikidata plus some of
the rule implicit in Wikidata through completion of Wikidata with some of these rules. As Wikidata
constantly changes due to edits these evaluations can not only be on a fixed, but current, version
of Wikidata but also against Wikidata subject to a current stream of changes, stressing the truth
maintenance aspect of rule systems.</p>
      <p>Although Wikidata is generally considered to be of good quality, there are errors in Wikidata. Another
evaluation task for rule-based systems on Wikidata is finding contradictions in Wikidata and then
ifnding likely causes for these contradictions.</p>
      <p>Wikidata currently sufers from not having a mechanism to show the implications of the intended
meaning of its constructs. A rule system that can efectively process Wikidata would provide a useful
service for anyone wanting to edit or use Wikidata and could end up being an oficial part of Wikidata.
This last benefit can work two ways—not only do rule systems benefit from use on Wikidata but
Wikidata can benefit from a better interface that can access the consequences of its data as determined
by the rule system. As we are interested in improving Wikidata we are willing to provide assistance to
any team that wants to use Wikidata to evaluated rule-based systems.</p>
      <p>What follows is a description of Wikidata, some rule sets that implement the intended meaning of
some parts of Wikidata, and more description of challenges and evaluations that might be performed
using Wikidata.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Wikidata</title>
      <p>
        Wikidata (https://wikidata.org) [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] is a large, freely-available knowledge graph built from the
contributions of many editors. As of September 2025, Wikidata contains more than 118 million items, such as
the item Q39246 (Richard Feynman),1 and over 10 thousand properties, such as P31 (instance of). Part
of the information about Q39246 is shown in Figure 1. Each item has a multi-lingual label and other
natural language information plus a collection of statements about the item. A statement can be thought
of, with some loss of precision, as a simple binary relationship, e.g. P31(Q39246,Q5) or instance
of(Richard Feynman,human); or as a more-complex relationship that contains qualifier pairs, e.g.,
P735(Q39246,"Richard",P1545,1) or given name(Richard Feynman,"Richard",series
ordinal,1). Statements also have a rank (preferred, regular, or deprecated) and can have references
supporting the statement. There were about 1.65 billion statements in Wikidata as of early 2025
(https://en.wikipedia.org/wiki/Wikidata) plus about the name number of facts that are not represented
as Wikidata statements, notably natural language labels and descriptions of items in multiple languages.
      </p>
      <p>The native form of Wikidata is a collection of Wikibase (https://www.mediawiki.org/wiki/Wikibase)
pages, that are variants of the MediaWiki (https://www.mediawiki.org/wiki/MediaWiki) pages used
in Wikipedias. There are translations of Wikidata into RDF, both as weekly dumps (https://dumps.
wikimedia.org/wikidatawiki/entities/) and as change logs.2 The “truthy” versions of these do not contain
qualifiers or references for statements and omit deprecated-rank statements and regular-rank statements
that contrast with preferred-rank statements. The “full” versions of these represent all the information
in Wikidata in RDF using a complex encoding that uses multiple RDF triples for each statement. There
are about 20 billion triples in the QLever endpoint (measured using a query in the endpoint itself). The
truthy dump is about one-tenth the size. Wikidata received between 300 and 1100 edits per minute
as of 2020 (https://wikitech.wikimedia.org/wiki/WMDE/Wikidata/Growth) with each edit potentially
changing multiple statements. This rate was expected to remain roughly constant.</p>
      <p>Wikidata is just a knowledge graph. There is no formal theory underlying Wikidata. There are
no rules nor any built-in inference process in Wikidata. What Wikidata does have is an intended
meaning for some of its properties and items. For example, Wikidata has two properties that are core to
its ontology: P31 (instance of) and P279 (subclass of). These two properties are considered to have
their usual meaning, i.e., subclass of is transitive and instances of classes are also instances of their
superclasses.
1In this paper entities from Wikidata will be often shown using their internal ID along with their English label.
2There is currently little public documentation on this interface, even though it is used to drive changes to the QLever SPARQL
Wikidata endpoint (https://qlever.cs.uni-freiburg.de/wikidata).</p>
      <p>Figure 1: Wikidata Page for Richard Feynman</p>
      <p>Wikidata does not provide any support for this intended meaning. Users wishing to access complete
subclass and instance relations generally use a SPARQL Wikidata endpoint and use SPARQL path
expressions (wdt:P279* and wdt:P31/wdt:P279*, respectively). This situation is by no means ideal
as users have to know to write this more-complex query, with the greater chance of getting it wrong.</p>
      <p>
        Partly as a result of the lack of a formal basis, there are errors in Wikidata. The presence of these
errors have been known for quite some time and there was recently an efort to characterize the errors
in the Wikidata ontology [
        <xref ref-type="bibr" rid="ref2 ref3 ref4">2, 3, 4</xref>
        ].
      </p>
      <p>
        This is not to say that there is nothing in Wikidata that can be used to point out errors in Wikidata.
Wikidata does have a constraint system (https://www.wikidata.org/wiki/Help:Property_constraints_
portal) and there are tools to evaluate the constraints against the data in Wikidata, producing reports
on constraint violations. We have recently developed tools to find violations of disjointness in Wikidata
[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] and problems with class order [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. These tools are not based on any formal description of Wikidata
and do not use any logic-based systems but instead are combinations of queries and procedural code.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Rules for Wikidata</title>
      <p>Given the lack of formal basis described above, how can Wikidata be used as a challenge or benchmark
for rule systems?</p>
      <p>What a rule system can do for Wikidata is to formalize and implement part of the intended meaning
of Wikidata—the meaning of Wikidata that is not just the raw data in Wikidata but encompasses the
informal meaning for quite a number of its items and, especially, properties.</p>
      <sec id="sec-3-1">
        <title>3.1. Basic Ontology Rules</title>
        <p>Perhaps the simplest use for rules3 in Wikidata is to implement the intended meaning of P31 (instance
of) and P279 (subclass of), with subclass being transitive and instances of classes being instances of
their superclasses.</p>
        <p>279(?, ?) ←
 31(?, ?) ←</p>
        <p>279(?, ?),  279(??)
 31(?, ?),  279(??)
(1)
(2)
These rules correspond to the SPARQL path expressions that Wikidata users employ when querying
the Wikidata ontology.</p>
        <p>This seems to be too simple for current rule systems. But there are still issues of size. With about 118
million items in Wikidata and 2 or 3 billion triples in truthy Wikidata expressed as RDF or 20 billion
triples in full Wikidata expressed as RDF even these simple rules could pose a worthwhile challenge
for rule systems. The transitive closure of subclass has about 125 million relationships and there are
about 1.5 billion total instance relationships as of September 2025 (determined by SPARQL queries)
indicating how many inferences would have to be computed to complete these basic ontology properties
in Wikidata.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Subproperty Rule</title>
        <p>A further complication is that the above properties are just regular properties, as opposed to being
part of the meta-language of Wikidata. So these properties can be afected by other parts of Wikidata,
notably P1647 (subproperty of). As there are subproperties of P31 (instance of) and also of P279
(subclass of), the normally-used queries for instance and subclass are incomplete. (To make ontology
SPARQL queries complete for a language that has subproperties is not possible in general.) Even for
Wikidata, which only has a few relevant subproperties, the required SPARQL queries for truly complete
answers for instance and subclass are quite complex.</p>
        <sec id="sec-3-2-1">
          <title>3We will use a neutral clause syntax for rules throughout this paper.</title>
          <p>The rule
?(?, ?) ←
 1647(?, ?) ←
 1647(?, ?), ?(?, ?)</p>
          <p>1647(?, ?),  1647(?, ?)
is needed to add the intended meaning of subproperty, namely that subproperty is transitive and
statements for a subproperty are also statements for its superproperties. The resultant set of rules is
no longer quite so simple, as it has a variable in predicate position. But even if the rule system does
not allow variables in predicate position, the simple trick of using a holds predicate can be employed,
resulting in a simple set of rules, albeit with a large number of consequences.</p>
        </sec>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Class Order</title>
        <p>Wikidata includes several classes with intended meaning that can be captured with rules.</p>
        <p>One group of such classes are the order classes: Q104086571 (first-order class), Q24017414
(secondorder class), Q24017465 (third-order class), and Q24027474 (fourth-order class). (There are other
order-related classes but they are not important here.)</p>
        <p>Class order is defined as follows: A class is first-order if all its instances are non-classes. A class is
 + 1st-order if all its instances are nth-order classes. Note that these rules should be taken to not mean
all instances in the current knowledge graph but instead as over all potential instances. So it is not
possible to compute class order just by looking at the current instances of a class.</p>
        <p>But it is possible to determine some axioms and rules for class order using the intended meaning of
the order classes above.</p>
        <p>(?, 1) ←  31(?, 104086571)
(?, 2) ←  31(?, 24017414)
(?, 3) ←  31(?, 24017465)
(?, 4) ←  31(?, 24027474)
(?, ?) ← (?, ?),  279(?, ?)
(?,  − 1) ← (?, ),  31(?, ?)
The rules introduce a new predicate for the order of a class, initially determined by it being an instance
of one of the order-inducing classes. Then order of a class is the same as order of any of its superclasses.
Also, the order of an item is one less than the order of any class it belongs to.</p>
        <p>
          These rules result in over 14 million first-order classes in Wikidata, over 2 million second-order
classes, and over 3 thousand third-order classes. (All these results were computed using SPARQL queries
that implement the above rules that were hand-crafted to some special circumstances in Wikidata [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ].)
Over ninety percent of the third-order classes are also second-order classes and ninety percent of the
second-order classes are also first-order classes.
        </p>
        <p>A non-empty class has at most one class order, which can be detected using the following rule:
(?) ← (?, ), (?, ),  ̸= 
So class order states that many Wikidata classes should be empty, as nearly all of these classes are not
empty in Wikidata the following rule results in many contradictions:
⊥ ←</p>
        <p>31(?, ?), (?)</p>
        <p>Problems with class order in Wikidata are real and do not just result from incorrect class order
statements. An interesting challenge for a rule system would be to determine potential root causes of
inconsistency involving class order to help reduce the issues in Wikidata.</p>
        <p>There are other inferences related to class order in Wikidata. Counting downward instance chains
from a class produces a minimum class order. Showing that an entity is both an instance and a subclass
of a class shows that the class cannot have an order. (It may be a variable order class, for example)
Combining these with the previous rules produces a theory of class order in Wikidata.
?(?, ?) ←</p>
        <p>1696(?, ?), ?(?, ?)
The latter is a kind of retributive characteristic, with rule:
?(?, ?) ←</p>
        <p>6609(?, ?), ?(?, ?), ?(?, ?)</p>
        <p>Although there are not very many Wikidata properties that belong to these classes or have values
for these properties, both P279 (subclass of) and P1647 (subproperty of) are instances of Q18647515
(transitive property) and P31 (instance of) P6609 (transitive over) P279 (subclass of).</p>
        <p>Implementing this part of basic ontology reasoning could be a significant challenge for rule systems.
(13)
(14)
(15)
(16)
(17)
(18)</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Property Characteristics</title>
        <p>Wikidata properties are entities and can belong to classes just like items can. Some of these classes have
intended meanings that are amenable to implementation via rules. These classes include Q18647518
(symmetric property), Q18647515 (transitive property), and Q18647519 (asymmetric property).
Currently, these intended meanings are implemented or checked by a combination of non-constraining
constraints4 and external programs (bots) that add and modify information in Wikidata.</p>
        <p>Wikidata properties can also have properties and some of these properties also have intended
meanings that are amenable to implementation via rules. These properties include P1696 (inverse
property) and P6609 (transitive over). The former is not complete inverse but instead is a partial
non-symmetric inverse, whose intended meaning is captured by the following rule:</p>
        <sec id="sec-3-4-1">
          <title>4Wikidata constraints mostly just add caution notes to Wikidata pages and produce reports.</title>
        </sec>
      </sec>
      <sec id="sec-3-5">
        <title>3.5. Disjointness</title>
        <p>Wikidata has the property P2738 (disjoint union of), which is a statement on a class that it is the disjoint
union of a list of other classes. This information is carried in a special qualifier: P11260 (list item).</p>
        <p>As a result, reasoning about disjointness in Wikidata requires information from the full dump of
Wikidata. But this qualifier only serves to allow disjointness to be more than a binary relationship so
reasoning about disjointness could be done in truthy Wikidata with just this special qualifier added,
possibly by translating disjoint union statements and their P11260 qualifiers into an -ary disjoint
union fact, but still using predicate P2738.</p>
        <p>With this translation, part of the intended meaning for disjoint unions consists of the following rules:
(?) ←
(?, ?) ←  2738(?, . . . , ?, . . . , ?, . . .)
(?, ?) ← (?, ?)
(?, ?) ← (?, ?),  279(?, ?)</p>
        <p>279(?, ?),  279(?, ?), (?, ?)
These rules first turn the Wikidata -ary disjoint union into binary disjoint statements. Then disjoint is
stated to be symmetric, subclasses of disjoint classes inherit disjointness, and a class that is the subclass
of two disjoint classes is an empty class.</p>
        <p>
          We conducted an analysis of disjointness of Wikidata in 2024 [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] and found that there were over 86
thousand classes that were necessarily empty because of disjointness and nearly 10 million entities
that were instances of two disjoint classes. (These disjointness problems were found by running many
queries against Wikidata and potentially missed a few problems because the queries used the usual way
of determining instance and subclass. A system based on rules would allow complete discovery.) Finding
disjointness problems adds more rules and more complexity to a Wikidata challenge or benchmark.
        </p>
      </sec>
      <sec id="sec-3-6">
        <title>3.6. Constraints</title>
        <p>As indicated above, Wikidata incorporates constraints (www.wikidata.org/wiki/Help:Property_
constraints_portal). A constraint in Wikidata is generally a non-enforcing condition that should
be met for relationships in a property.</p>
        <p>Most of the constraints are very local, such as the single-value constraint, which checks that each
item has only one value for a property, such as the single-value constraint on P22 (father).</p>
        <p>Constraints can mostly be checked directly in the Wikidata data structures, as they involve
properties that do not have subproperties or have any special intended meaning and only look for simple
characteristics of the property and its values on an item.</p>
        <p>But complete checking of some constraints involves properties that do have subproperties or special
intended meaning, such as the constraints on P31 (instance of). Constraints on these properties are
incompletely handled within Wikidata. (Some constraints are handled by external bots—programs
outside the Wikidata software that are run periodically and either produce constraint violation reports
or modify Wikidata in some way to satisfy constraints.)</p>
        <p>
          Wikidata constraints are described using natural language and do not have a formalization in Wikidata,
but have been formalized in SHACL by Ferranti et al [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]. Expressing them as rules is fairly simple.
        </p>
        <p>As there are many constraints in Wikidata and they cover many relationships, another challenge
problem for rules would thus be to determine constraint violations in Wikidata. There are many of
these, as the voluminous constraint violations reports (www.wikidata.org/wiki/Wikidata:Database_
reports/Constraint_violations) show.5</p>
        <p>A rule system that could efectively process the constraints of Wikidata would provide a useful
service for Wikidata.</p>
      </sec>
      <sec id="sec-3-7">
        <title>3.7. Qualifiers</title>
        <p>One of the challenges of Wikidata as far as rules go is how to incorporate qualifiers in general.</p>
        <p>
          Qualifiers can add information to a statement, such as the disjoint components above. But qualifiers
can also contextualize, or limit, a statement [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ], for example by using temporal qualifiers, including
P580 (start time) and P582 (end time). Aljalabout et al. [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] categorize the major qualifiers and propose
using a many-sorted logic in which to capture their meaning.
        </p>
        <p>
          There are several challenges for rules here—a diferent kind of challenge than one for rule systems.
First, to come up with a base logic for qualifiers, such as Attributed Description Logics [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ], and a
syntax for rules there. Second, to either build a theory in that logic or an extension of the logic to handle
contextualizing qualifiers. Third, to produce rules that formalize the meaning of important qualifiers in
Wikidata.
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Incorrect Information in Wikidata</title>
      <p>As the above discussion indicates, Wikidata has many contradictions or potential contradictions when
its intended meaning is considered. Many of these contradictions are hard to detect without some sort
of inferencing. Helping to reduce inconsistencies in Wikidata is thus a useful challenge for rule systems.</p>
      <p>It can make sense to assume that some parts of a dataset are consistent and suitable for inference.
Even though Wikidata data is generally of high quality, and most statements in it are correct, having
even a low percentage of incorrect information can result in many inferences that are not correct. For
example, one might want to assume that instance information in Wikidata is correct enough to find
problems with class order or assume that class order information is correct enough to find problems
with instance. But both have errors and it is unclear how to use one sort of information to find problems
with the other.</p>
      <p>An interesting challenge is thus to better use the information in Wikidata, potentially along with
external information, to detect problems and potential fixes. The role of the rule system in this challenge
5These reports miss some constraint violations.
is to detect problems, as evidenced by inferred contradictions, and come up with potential causes of
the problems. The overall system would provide guidance, by comparing with other good information
sources, by using information in Wikidata to estimate information quality, or by other means.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Reasoning Modes</title>
      <p>There are several modes of operation that a rule system can employ. A rule system can be
forwardchaining, where inferences are performed and their consequences added to a knowledge base and
retained indefinitely in advance of any requests for information. A rule system can be
backwardchaining, where inferences are performed only as necessary to answer requests and not retained after
the request has been answered. A rule system can eschew these pure modes by sometimes adding
consequences that are not needed to answer requests or remembering consequences of past inferences
for future use.</p>
      <p>A challenge for forward-chaining rule systems is simply to compute the consequences of part of the
intended meaning of Wikidata, using rules that formalize this intended meaning.</p>
      <p>For backward-chaining or mixed rule systems a challenge needs to also include a set of queries. Instead
of using a fixed set of queries, a challenge or benchmark could instead be based on parameterized
queries with parameter values determined by sampling from the current version of Wikidata. For
example, a very simple parameterized query would be to determine the full set of types of an item, as in
⊥ ←
 31(?, ?)
(19)
where ?p is a parameter, not a variable. Parameter values can be determined by various means such as
sampling from the results of SPARQL queries, for example</p>
      <p>SELECT ?p WHERE { ?p wdt:P31 wd:Q2424752}
to sample from Q2424752 (product).</p>
      <sec id="sec-5-1">
        <title>5.1. Updates</title>
        <p>One issue with a rule system for Wikidata is that Wikidata is edited frequently. A rule system for
Wikidata that records consequences needs to be able to forget these consequences if their support is
removed. As Wikidata is updated many times per minute, responding to edits can in itself be a challenge
for a rule system.</p>
        <p>So an additional challenge for any rule system that records consequences is to keep up with the edits
to Wikidata. At the same time that the rule system is processing edits it also needs to be able to process
queries, so the update process needs to not consume too many compute resources.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Running a Wikidata Challenge or Benchmark</title>
      <p>Wikidata data is available under the CC0 license (creativecommons.org/publicdomain/zero/1.0/), which
permits free use for any purpose, so there is no problem in using Wikidata for evaluating rule systems.
The current dumps of Wikidata are large but readily available and their format is well documented, as
described at www.wikidata.org/wiki/Wikidata:Database_download. These dumps are only retained for
a short period of time, so any reuse of historical versions of Wikidata would require placing the dump
in a separate repository. The stream of edits to Wikidata was mostly used internally and is only now
being made available for external use. Documentation is being prepared for this stream.</p>
      <p>Aside from the size of Wikidata, running a challenge or benchmark for living Wikidata would require
more work because of the need to keep previous results updated. When a Wikidata-related problem is
ifrst proposed, instead of providing a fixed dataset there will have to be a process for constructing the
dataset from the current version of Wikidata. Sometimes this might be just obtaining a current RDF or
other dump of Wikidata and possibly a current stream of updates but sometimes some pre-processing
might also be needed. For problems that need a set of queries or other accesses, there also needs to be a
process for constructing these queries that uses the current version of Wikidata. Finally, there needs to
be a process for running the challenge or benchmark that can be repeated by others and with diferent
rule systems. Some of these requirements are already in place to support repeatability, the processes
need to work for not just a given dataset and query set but for new versions of Wikidata.</p>
      <p>Running the challenge or benchmark later will not only involve creating the current dataset and
query set and running it on a new rule system but also updating the results for other rule systems on
the current data. As well, a process for running the new rule system on future versions of the data
needs to be created.</p>
      <p>A challenge or benchmark for updates also needs means to receive the update stream, process updates
as necessary, and send the update stream to the rule system. There also needs to be a way of also
sending queries to the rule system during updates.</p>
      <p>Repeatability can be supported as it always has been by requiring evaluations to include the actual
data that they used for their evaluations. But to achieve prominence, any rule system has to perform
well not only at a single point but needs to continue to perform well as Wikidata evolves. This prevents
lock-in to the single dataset and the common situation of marginal improvements on that dataset.</p>
      <p>Our ultimate goal in proposing Wikidata challenges is to improve Wikidata itself, so we would like
to evaluate systems based on how well they might support this goal. But that by no means an objective
measure, so we must determine which metrics best approximate what the community would deem
useful. We thus feel that correctness and the ability to perform the task are the two primary evaluation
metrics, making time a secondary metric. On the correctness side, an unsupported inference would be
disqualifying. For evaluations involving updates the ability to keep up with the update stream becomes
a primary metric.</p>
    </sec>
    <sec id="sec-7">
      <title>7. Conclusion</title>
      <p>Using a challenge or benchmark based on parts of the intended meaning of living Wikidata has benefits
for rule systems.</p>
      <p>Wikidata can be viewed in several ways as a challenge or benchmark. Wikidata has over 118 million
entities. There is truthy Wikidata, where Wikidata is viewed as only having standard facts, including
Wikdiata statements, in a standard Boolean logic. Even in truthy Wikidata there is the need to either
allow variables in rules or to encode facts to get around this need. There is full Wikidata, where Wikidata
statements have associated qualifiers, requiring the ability to work outside of simple logics, such as a
temporal logic, or encoding these extensions in a simple logic. Truthy Wikidata has over 2 billion facts
and full Wikidata has a order of magnitude more. The intended meaning of entities in Wikidata give
rise to a variety of rulesets, ranging from small rulesets in simple logics, but with large consequence
sets, to large rulesets on temporal and other logics. Wikidata is updated hundreds of times per minute.
All of these place stresses on performance characteristics of rule systems, so even fixed snapshots of
Wikidata are a source of good challenges and benchmarks for rule systems. As Wikidata is continually
updated, it can serve as a living challenge or benchmark, preventing lock-in to a fixed dataset.</p>
      <p>But the best argument for using living Wikidata as a challenge problem for rule systems is that
Wikidata does not provide good support for the intended meaning of even entities that are central to it
and its ontology. This places undue burdens on users of Wikidata and creates an opening for systems
that can efectively implement this intended meaning. A rule system that can reliably and eficiently
handle Wikidata could end up being part of the software that implements Wikidata and thus used and
potentially seen by the large community of Wikidata editors.</p>
      <p>We would propose three initial challenges involving Wikidata for rule systems. The first challenge is
to complete the truthy Wikidata ontology based on the intended meainings of P31 (instance of), P279
(subclass of), and P1647 (subproperty of). This challenge is simple and likely doable by multiple rule
systems in reasonable time. The second challenge is to complete the full Wikidata ontology based on
at least all the intended meanings above except constraints, and calculating temporal extents for the
inferred statements. This challenge requires rule systems that can handle qualified statements, and
may spur development of such systems. These two challenges can be run as both static and updating
variants. The third challenge is to find good changes that can fix problems related to disjointness or
class order or constraint violations. This last challenge would be judged by members of the Wikidata
community.</p>
    </sec>
    <sec id="sec-8">
      <title>Declaration on Generative AI</title>
      <p>When preparing this paper for publication, one of the authors did numerous Google web searches to try
to find out how to include the correct fonts and copyright clause. The results of these searches included
results from a generative AI system. Some of these generative AI results were skimmed or read by the
author, but none of the information in them was useful.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>D.</given-names>
            <surname>Vrandečić</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Krötzsch</surname>
          </string-name>
          ,
          <article-title>Wikidata: A free collaborative knowledgebase</article-title>
          ,
          <source>Communications of the ACM</source>
          <volume>57</volume>
          (
          <year>2014</year>
          )
          <fpage>78</fpage>
          -
          <lpage>85</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>D.</given-names>
            <surname>Ammalainen</surname>
          </string-name>
          ,
          <article-title>Wikidata ontology issues: Suggestions for prioritisation based on the perceived frequency of occurrence and the severity of impact on data re-use, commons</article-title>
          .wikimedia.org/ wiki/File:Wikidata_ontology_issues_%E2%
          <volume>80</volume>
          %94_suggestions_for_prioritisation_
          <year>2023</year>
          .pdf,
          <year>2023</year>
          . Accessed 2025-
          <volume>05</volume>
          -12.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>M.</given-names>
            <surname>Abdulai</surname>
          </string-name>
          , L. Lacroix, Wikidata:
          <article-title>Ontology issues prioritization, www</article-title>
          .wikidata.org/wiki/Wikidata: Ontology_issues_prioritization,
          <year>2023</year>
          . Accessed 2025-
          <volume>05</volume>
          -12.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>L.</given-names>
            <surname>Pintscher</surname>
          </string-name>
          ,
          <article-title>Wikidata survey on ontology issues, potential solutions, www</article-title>
          .wikidata.org/wiki/ Wikidata_talk:Ontology_issues_prioritization#Overview_of_potential_solutions,
          <year>2023</year>
          . Accessed 2025-
          <volume>02</volume>
          -17.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>E. A.</given-names>
            <surname>Doğan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. F.</given-names>
            <surname>Patel-Schneider</surname>
          </string-name>
          ,
          <article-title>Disjointness violations in Wikidata</article-title>
          , in: Fifth
          <source>International Knowledge Graph and Semantic Web Conference</source>
          ,
          <year>2024</year>
          .
          <article-title>A longer version available as arxiv</article-title>
          .
          <source>org/ abs/2410</source>
          .13707.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>P. F.</given-names>
            <surname>Patel-Schneider</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. A.</given-names>
            <surname>Doğan</surname>
          </string-name>
          ,
          <article-title>Class order disorder in Wikidata and first fixes, arxiv</article-title>
          .org/abs/ 2411.15550,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>N.</given-names>
            <surname>Ferranti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Polleres</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. F. de Souza</surname>
          </string-name>
          , S. Ahmetaj,
          <article-title>Formalizing property constraints in wikidata</article-title>
          ,
          <source>in: Wikidata'22: Wikidata workshop at ISWSC</source>
          <year>2022</year>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>P. F.</given-names>
            <surname>Patel-Schneider</surname>
          </string-name>
          ,
          <article-title>Contextualization via qualifiers</article-title>
          ,
          <source>in: First International Workshop on Contextualized Knowledge Graphs at ISWC</source>
          <year>2018</year>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>S.</given-names>
            <surname>Aljalbout</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Falquet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Buchs</surname>
          </string-name>
          ,
          <article-title>Handling Wikidata qualifiers in reasoning, arxiv</article-title>
          .org/abs/2304. 03375,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>M.</given-names>
            <surname>Krötzsch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Marx</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ozaki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Thos</surname>
          </string-name>
          ,
          <article-title>Attributed description logics: Ontologies for knowledge graphs</article-title>
          ,
          <source>in: The 16th Internaional Semantic Web Conference</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>