=Paper=
{{Paper
|id=Vol-373/paper-9
|storemode=property
|title=The Annual SUMO Reasoning Prizes at CASC
|pdfUrl=https://ceur-ws.org/Vol-373/paper-05.pdf
|volume=Vol-373
|dblpUrl=https://dblp.org/rec/conf/cade/PeaseSST08
}}
==The Annual SUMO Reasoning Prizes at CASC==
The Annual SUMO Reasoning Prizes at CASC
Adam Pease1, Geoff Sutcliffe2, Nick Siegel1, Steven Trac2
1
Articulate Software
apease|nsiegel[at]articulatesoftware.com
2
University of Miami
geoff|strac[at]cs.miami.edu
Abstract
Previous CASC competitions have focused on proving difficult problems on small numbers of axioms.
However, typical reasoning applications for expert systems rely on knowledge bases that have large numbers
of axioms of which only a small number may be relevant to any given query. We have created a category in
the new LTB division of CASC to test this sort of situation. We present an analysis of performance of last
year's entrants in CASC to show how they perform before any opportunity for tuning them to this new
competition.
1. Introduction
Previous CASC competitions have focused on proving difficult problems on relatively small numbers of
axioms. However, typical reasoning applications for expert systems rely on knowledge bases that have
large numbers of axioms, of which only a small number may be relevant to any given query. We have
chosen the Suggested Upper Merged Ontology as the basis for a category of the new Large Theory
Batch (LTB) division of CASC.
The Suggested Upper Merged Ontology (SUMO) (Niles & Pease, 2001) is a free, formal
ontology of about 1000 terms and 4000 definitional statements. It is provided in the SUOKIF language
(Pease, 2003), which is a first order logic with some secondorder extensions, and also translated into
the OWL semantic web language (which is a necessarily lossy translation, given the limited
expressiveness of OWL). In prior work we have described how we transformed SUMO into a strictly
firstorder form (Pease&Sutcliffe, 2007). SUMO has also been extended with a MIdLevel Ontology
(MILO), and a number of domain ontologies, which together number some 20,000 terms and 70,000
axioms. SUMO has been mapped to the WordNet lexicon (Fellbaum, 1998) of over 100,000 noun, verb,
adjective, and adverb word senses (Niles & Pease, 2003), which not only acts as a check on coverage
and completeness, but also provides a basis for work in natural language processing (Pease & Murray,
2003) (Elkateb et al, 2006) (Scheffczyk et al, 2006). SUMO is now in its 75th free version; having
undergone five years of development, review by a community of hundreds of people, and application in
expert reasoning and linguistics. Various versions of SUMO have been subjected to formal verification
with Vampire (Riazanov&Voronkov 2002), which until recently was the only prover we had integrated
into our browsing and inference tool suite called Sigma (Pease, 2003). SUMO and all the associated
tools and products are available at www.ontologyportal.org.
2.The Competition
The SUMO inference prizes totaling US$3000.00 will be awarded to the best performance on the SMO
category of the LTB division of CASC, held at IJCAR 2008. The LTB division has an assurance
ranking class and a proof ranking class. In each ranking class the winner will receive $750, the second
place $500, and the third place $250 (a system that wins the proof ranking class might also win the
assurance ranking class).
66
We created an additional test to support the participation of modelfinders. The SUMO
validation prize totaling US$300 will test these systems, and hopefully improve SUMO by finding any
problems with the theory. Three subdivisions, each with a $100 prize will be given to those systems
which
1. Verify the consistency of, or provide feedback to repair, the base SUMO ontology.
2. Verify the consistency of, or provide feedback to repair, the combined SUMO and MILO ontologies.
3. Verify the consistency of, or provide feedback to repair, the combined SUMO, MILO, and domain
ontologies.
The winners of the SUMO challenges will be announced and receive their awards at IJCAR following
successful completion of a challenge.
3.Example Test
To give a flavor of what the tests consist of, we present one of them. The question posed to the system
can be described as “Can a human perform an intentional action if he or she is dead?”. We create in the
test an example instance of an action
(instance DoingSomething4-1 IntentionalProcess)
then state that an individual is performing the action
(agent DoingSomething4-1 Entity4-1)
and that the individual is human
(instance Entity4-1 Human)
The successful theorem prover will then find the following axioms and apply them to prove the
conjecture
(<=>
(instance ?X4 Agent)
(exists (?X5)
(agent ?X5 ?X4)))
(subclass IntentionalProcess Process)
(=>
(and
(subclass ?X403 ?X404)
(instance ?X405 ?X403))
(instance ?X405 ?X404))
(=>
(and
(agent ?X5 ?X4)
(instance ?X5 IntentionalProcess))
(and
(instance ?X4 CognitiveAgent)
(not
(holdsDuring
(WhenFn ?X5)
(attribute ?X4 Dead)))))
67
We should note that this proof has the interesting feature that although the form appears to be second
order (holdsDuring arg ), the system treats the embedded formula as an uninterpreted
list and is able to solve the problem simply by unifying clauses in the list.
While this example is trivial when the necessary axioms are found ahead of time, it becomes
very challenging in the context of a large knowledge base, where, in a practical situation, the relevant
axioms cannot be known ahead of time. There are hundreds or thousands of axioms involving the term
“agent” in SUMO, for example, and the successful theorem prover will have to hunt through those
axioms very quickly in order to find just the ones that are relevant to the query being posed.
4.Analysis
In order to test whether the competition was even reasonable, we decided to run it on all the provers in
the SystemOnTPTP suite. These were Bliksem 1.12, CARINE 0.734, CiME 2.01, Darwin 1.4.1,
DarwinFM 1.4.1, DCTP 1.31, E 0.999, EKRHyper 1.0, EQP 0.9d, Equinox 1.3, Fampire 1.3, Faust 1.0,
FDP 0.9.16, Fiesta 2, Gandalf c2.6, Geo 2007f, GrAnDe 1.1, iProver 0.2, leanCoP 2.0, LeanTAP 2.3,
Mace2 2.2, Mace4 1207, Matita 0.1.0, Metis 2.0, Muscadet 2.7a, Otter 3.3, Paradox 2.3, Prover9 1207,
SSETHEO 0.0, SETHEO 3.3, SNARK 20070805, SOS 2.0, SPASS 3.0, SRASS 0.1, Theo 2006,
Vampire 9.0, Waldmeister 806, zChaff 04.11.15, Zenon 0.5.0. We gave each prover 600 seconds on
each of 102 problems, generated from 33 distinct queries (possibly with some additional assertions to
the knowledge base) each tested with just the ~4000 axioms in SUMO, the ~9000 axioms of
SUMO+MILO or the tens of thousands of axioms in SUMO+MILO and all the domain ontologies.
Overall SUMO SUMO+MILO All
Vampire 9.0 Vampire 9.0 Vampire 9.0 Metis 2.0
Metis 2.0 E 0.999 Metis 2.0 Zenon 0.5.0
E 0.999 iProver 0.2 SNARK 20070805 Equinox 1.3
iProver 0.2 leanCoP 2.0 Zenon 0.5.0
leanCoP 2.0 Metis 2.0 Equinox 1.3
Darwin 1.4.1 Darwin 1.4.1 Muscadet 2.7a
Zenon 0.5.0 Fampire 1.3
Equinox 1.3 SNARK 20070805
Fampire 1.3 Zenon 0. 5.0
SNARK 20070805 Equinox 1.3
Muscadet 2.7a Muscadet 2.7a
SPASS 3.0 SPASS 3.0
Faust 1.0 Faust 1.0
Table 1: Performance ranking
Overall performance is shown in the first column above with Vampire achieving first place. All
other provers not listed failed to solve any of the problems. The best performance with SUMO alone is
shown then SUMO+MILO and finally performance with all the domain ontologies loaded. The best
performing provers still did not solve a majority of the 105 problems in the test set. Vampire solved 31,
Fampire 20, E 15 and Metis 14, with the other provers in the single digits or no solutions at all. Prover
failing to find solutions were stopped generally because of timeouts, rather than errors in parsing or
memory space. Average running times approached the 600 seconds allocated for all provers because of
68
the low percentage of solved problems.
The differing strengths of several of the provers suggested creating a “metaprover” combining
several systems. The strategy is to give Vampire 400 seconds, then give Metis up to 200 seconds if
Vampire failed to find a proof. The combined system gets 33 answers compared to 31 for Vampire alone
or 14 for Metis alone, and performance overall is slightly better at 48158 seconds vs. 55419 for Metis
and 48599 for Vampire. We might be able to tweak the timeslice allocation to do still better, although
further efforts in that regard could be considered overtraining to this particular problem set.
We performed an analysis to determine what set of systems would cover the maximum number
of problems (see Figure 1). This is termed a “SOTA” analysis as per (Sutcliffe & Suttner 2001).
Vampire solved eight problems solved by no other prover. Metis uniquely solved two, and Fampire 1.
This analysis suggests that we should revisit creation of a metaprover composed of Vampire, Metis and
Fampire.
Fampire Vampire Metis
iProver Equinox
E
SPASS Zenon
leanCoP SNARK Muscadet
Faust
Figure 1: SOTA analysis
5.Conclusions
We have created a category called “SMO” in the new LTB division of CASC to motivate high
performance reasoning on practical problems using a broad knowledge base. We believe this will yield
some exciting research results, as well as provide the application development community with provers
that are more closely optimized to the needs to one sort of practical inference. We have run the tests
with existing theorem provers and found the competition to be a reasonable goal for these systems.
With tuning, we expect even better performance.
In the future we expect to expand the number of tests in the SMO category. We also anticipate
providing a “stratified” set of tests of different expressiveness, in which we extract the horn clause and
description logic subsets of SUMO and provide tests on those subsets.
69
Acknowledgments
This work has been funded by a number of sources, including the Army Research Institute. We are grateful for
their investment.
References
Elkateb, S., Black, W., Rodriguez, H, Alkhalifa, M., Vossen, P., Pease, A. and Fellbaum, C., (2006). Building a WordNet for
Arabic, in Proceedings of The fifth international conference on Language Resources and Evaluation (LREC 2006).
Fellbaum, C. (ed. ) WordNet: An Electronic Lexical Database. MIT Press, 1998.
Genesereth, M., (1991). “Knowledge Interchange Format'', In Proceedings of the Second International Conference on the
Principles of Knowledge Representation and Reasoning, Allen, J., Fikes, R., Sandewall, E. (eds), Morgan Kaufman
Publishers, pp 238249.
Hayes, P., and Menzel, C., (2001). A Semantics for Knowledge Interchange Format, in Working Notes of the IJCAI2001
Workshop on the IEEE Standard Upper Ontology.
Niles, I & Pease A., (2001). “Towards A Standard Upper Ontology.” In Proceedings of Formal Ontology in Information
Systems (FOIS 2001), October 1719, Ogunquit, Maine, USA, pp 29. See also http://www.ontologyportal.org
Niles, I., and Pease, A. (2003) Linking Lexicons and Ontologies: Mapping WordNet to the Suggested Upper Merged
Ontology, Proceedings of the IEEE International Conference on Information and Knowledge Engineering, pp
412416.
Pease, A., (2003). The Sigma Ontology Development Environment, in Working Notes of the IJCAI2003 Workshop on
Ontology and Distributed Systems. Volume 71 of CEUR Workshop Proceeding series. See also
http://sigmakee.sourceforge.net
Pease, A., (2004). Standard Upper Ontology Knowledge Interchange Format. Unpublished language manual. Available at
http://sigmakee.sourceforge.net/
Pease, A., and Murray, W., (2003). An English to Logic Translator for Ontologybased Knowledge Representation
Languages. In Proceedings of the 2003 IEEE International Conference on Natural Language Processing and
Knowledge Engineering, Beijing, China, pp 777783.
Pease, A., and Sutcliffe, G., (2007) First Order Reasoning on a Large Ontology, in Proceedings of the CADE21 workshop
on Empirically Successful Automated Reasoning on Large Theories (ESARLT).
Ramachandran, D., P. Reagan, K. Goolsbey. FirstOrderized ResearchCyc: Expressivity and Efficiency in a CommonSense
Ontology. In Papers from the AAAI Workshop on Contexts and Ontologies: Theory, Practice and Applications.
Pittsburgh, Pennsylvania, July 2005.
Riazanov A., Voronkov A. (2002). The Design and Implementation of Vampire. AI Communications, 15(23), pp. 91—110.
Scheffczyk, J., Pease, A., Ellsworth, M., (2006). Linking FrameNet to the Suggested Upper Merged Ontology, in
Proceedings of Formal Ontology in Information Systems (FOIS2006), B. Bennett and C. Fellbaum, eds, IOS Press,
pp 289300.
Sutcliffe G., Suttner C.B. (1998), The TPTP Problem Library: CNF Release v1.2.1, Journal of Automated Reasoning 21(2),
177203.
Sutcliffe G., Suttner C.B. (2001), Evaluating General Purpose Automated Theorem Proving Systems, Journal of Artificial
Intelligence 131(12), 3954.
70