=Paper=
{{Paper
|id=None
|storemode=property
|title=The Limitations of Description Logic for Mathematical Ontologies: An Example on Neural Networks
|pdfUrl=https://ceur-ws.org/Vol-938/ontobras-most2012_paper7.pdf
|volume=Vol-938
|dblpUrl=https://dblp.org/rec/conf/ontobras/FreitasL12
}}
==The Limitations of Description Logic for Mathematical Ontologies: An Example on Neural Networks==
The Limitations of Description Logic for Mathematical Ontologies: An Example on Neural Networks Fred Freitas, Fernando Lins Informatics Center - Federal University of Pernambuco (CIn - UFPE) Av. Prof. Luis Freire, s/n, Cidade Universitária, 50740-540, Recife – PE – Brazil {fred,fval}@cin.ufpe.br Abstract. In this work, we discuss appropriate formalisms to account for the representation of mathematical ontologies, and, particularly, the problems for employing Description Logics (DL) for the task.DL has proven to be not expressive enough to such tasks, because it cannot represent, for instance, simple numerical constraints, except of cardinality constraints over instances of individuals and relation instances. Therefore, for such representations, we advocate the use of a more expressive formalism, based at least on first-order logic using a top ontology as support vocabulary. We also provide a thorough example on the representation of an Artificial Neural Network (ANN) defined in KIF (Knowledge Interchange Format), along with a discussion on the requirements needed for such a representation and the results achieved. 1 Introduction All In this work, we introduce a discussion on the appropriate formalism to carry out the task of representing mathematical knowledge, and particularly the problems for doing it with Description Logics (BAADER ET AL, 2003). It departed from an attempt to develop an Ontology of Multilayer Perceptron (MLP) Artificial Neural Networks (ANN). Our initial intent was to define it using the Semantic Web standardized description logic language OWL (Ontology Web Language) (WELTY ET AL, 2004), which would offer us a lot of benefits, like mature development tools, availability of reasoners, large community of users, etc. On creating such ontology, our work was focused in developing a conceptual model that would include the necessary mathematics to enable the future development of knowledge-based applications that can reason over artificial neural networks, such as an intelligent agent capable of interacting with users to answer questions about neural networks that can even create and run them. During the development of the ontology, we faced several representational problems while trying to represent the mathematical expressions inherent to neural networks in Description Logic (DL). Since this formalism is based on set theoretic semantics, it fits properly to representing domains in which relationships among sets suffice. For such domains, DL terminologies can accurately describe the domains’ classes. On the one hand, the choice of this formalism for the Semantic Web lies actually on these grounds. On the other hand, it is not endowed with any apparatus to deal with many types of mathematical operations and constraints that would turn it into a more generic formalism, like what can be represented by Prolog, for instance. Moreover, a core mathematical ontology is required as a vocabulary to make for the needed complex well-defined mathematical meanings. 84 Therefore, for such representations, we advocate the use of a more expressive formalism, based at least on full first-order logic, and using a rich, well-defined support vocabulary. In order to illustrate the discussion, we first introduce the required basic mathematical definitions about MLP neural networks and the expressiveness requirements that a language should fulfill to represent the ontology. Then, we discuss the practical expressiveness limitations that hamper the use of Description Logic for such representations. Finally, we present our MLP ANN ontology and some of its more important definitions. We developed it in the Standard Upper Ontology Knowledge Interchange Format (SUO-KIF) language (PEASE 2009) and used the top ontology Suggested Upper Merged Ontology (SUMO) (NILES & PEASE, 2001) as support vocabulary, which proved well-tailored for our mathematical representations. We complete the article giving a very brief introduction of the SUO-KIF language and the top ontology SUMO, discussing some related work, future work and conclusions.Full papers must respect the page limits defined by the conference. Conferences that publish just abstracts ask for one-page texts. 2 An Example: Representing the MLP ANN in an Ontology Artificial Neural Networks (ANNs) were developed as a rough abstraction of biological neural networks. An ANN consists of a number of nodes that perform simple numerical processing. Its nodes are highly interlinked and grouped in layers. Mathematically speaking, it implements complex function approximators (HAYKIN, 1994). ANNs are used in the computer mainstream to perform pattern learning and recognition. The most commonly used ANN is the Multilayer Perceptron (MLP) (HAYKIN, 1994). It has three layers (entry, hidden and output) or more (when more than one hidden layer is used). Each layer consists of a set of “artificial neurons” connected with the neighbor layers, as displayed in the figure 1 below. Figure 1− A MLP Artificial Neural Network, its layers and neurons. A MLP ANN “learns” the implicit function it should approximate by computing, correcting and propagating classification errors throughout all its neurons, as follows. In the training phase, the MLP ANN is given a set of problem instances, usually represented as numerical values, along with their solutions, and process the data many times to assign correctly the neurons’ parameters. Synapses among artificial neurons are modeled via activation functions, which trigger the delivery of numerical stimuli as inputs to the next layer’s neurons. At the end of the learning phase, the network is tuned with the parameters so as to reflect as closely as possible the classification function implicit in the training set. 85 The Backpropagation algorithm (HAYKIN, 1994) is used to train the MLP ANN. It is divided into two phases. In the forward phase, the network is run and a classification output is calculated. In the backward phase, it compares the output with the correct result and, if wrong, it calculates and propagates the error through the weights of the connections among all the neurons of the network. Therefore, the learning ability of MLP ANNs lies on the neurons. The algorithm iterates between these two phases until it reaches a stopping criteria. These two phases are explained in detail in the next subsections. 2.1 Forward Phase Here, the ANN basically computes summations of a product. At first, each neuron receives its entry and propagates it to the first hidden layer’s neurons, if its activation function “triggers”. 𝑝 For each neuron of the hidden layer, the potential of activation of the neuron j (𝑛𝑒𝑡𝑗 , 𝑖𝑛 𝑡ℎ𝑒 𝑓𝑜𝑟𝑚𝑢𝑙𝑎) is calculated by summating all synapses weights (𝑤𝑗𝑖 ) from neurons i that reach neuron j, multiplied by the outputs of neurons ii: 𝑛 𝑛𝑒𝑡𝑗𝑝 = � 𝑥𝑖𝑝 𝑤𝑗𝑖 (1) 𝑖= 1 Then, the activation potential is applied to an activation function that will return the output of the neuron. One of the most used activation functions is the logistic function (also called sigmoid), that can be seen below: 1 𝑃�𝑛𝑒𝑡𝑗𝑝 � = 𝑝 (2) −𝑛𝑒𝑡𝑗 1+𝑒 The output of the neuron, given by the result of the activation function, typically consists of a real value between -1 and 1. A synapse is defined as the propagation of the output as an entry to the next layer, or, if it is the last layer, one of the outputs belonging to the output vector of the network. Synapses in a neuron occur if its output is higher than a given threshold, usually set to 0. Synapses are then propagated (or not) through the network until they reach the output layer, finishing the forward phase. This phase is executed when the network is in the execution mode too. 2.2 Backward Phase The first step of this phase consists in calculating the error between the output of the neuron j from the forward phase and the expected output from the instance of the training set, as in the following formula 𝛿𝑗 = �𝑑𝑗 − 𝑦𝑗 �𝑓 ′ (𝑛𝑒𝑡𝑗 ) (3) where 𝑑𝑗 is the desired output, 𝑦𝑗 the actual output, and 𝑓 ′(𝑛𝑒𝑡𝑗 ) the activation function of the neuron’s derivate. Hidden nodes’ errors are computed as follows: 𝑛 ′ 𝛿𝑗 = 𝑓 (𝑛𝑒𝑡𝑗 ) � 𝛿𝑙 𝑤𝑙𝑗 (4) 𝑖= 1 86 They are the product of the activation function derivate of the current neuron and the summation of the products between the error (𝛿𝑗 ) and each of the last layers’ weights (𝑤𝑙𝑗 ). Finally, the synapse weights that connect current neuron j with its antecessors (neurons i) are given by: 𝑤𝑗𝑖 (𝑡 + 1) = 𝑤𝑗𝑖 (𝑡) + 𝜂𝛿𝑗 (𝑡)𝑥𝑖 (𝑡) (5) where 𝑤𝑗𝑖 (𝑡) stands for the old weight, 𝑤𝑗𝑖 (𝑡 + 1) means the new weight, 𝜂 is the learning rate that regulates learning speed, 𝛿𝑗 (𝑡) is the error and 𝑥𝑖 (𝑡) is the input signal of the neuron with the old weight. The rough idea is slowly (as tuned by 𝜂) distributing the error issued by the network through all the layers and neurons. In the next section, we enlist the requirements for representing such knowledge. 3 Requirements of Candidate Formalisms and Languages To address the problem of choosing the knowledge representation formalism and language to represent the MLP ANN, we had to consider a number of representation requirements. We needed enough expressive power to represent the mathematical concepts on the ontology. Potential candidate formalisms and languages should, at the same time, allow us to do it in an easy way, as well as give us the possibility of computing results afterwards. Below, we list the most important requirements for that: • Ability to represent numerical constraints - as seen above, MLP encompasses plenty of calculations. The ontology must be capable of representing them in an straight-forward way, as well as the constraints over them involving mathematical operations like summations, exponentiations or function derivations. • Availability of a rich vocabulary of abstract definitions - the MLP ANN Ontology makes use of a lot of abstract concepts that are fulfilled either by a core mathematical ontology or by a top ontology which must include mathematical expressions, set theory, graph theory, algorithms, sequences, etc. • Ability to represent n-ary relations - the freedom to use unlimited arities in relations leaves ontologists free to use some advanced constructs. For instance, a synapse connects two neurons, thus possessing a ternary relation named connects that has, as arguments, the synapse and the two neurons. • Ability to represent functions - when dealing with mathematics concepts, we often need to represent functions. In the MLP ANN, we had to represent the activation function and the stopping criteria based on functions. In the next section, we discuss the difficulties in fulfilling such criteria with DL. 4 Description Logics’ Representational Problems regarding Mathematical Ontologies DL and its Ontology Web Language OWL present in practice a number of limitations in expressiveness that prevented us from using it for representing the knowledge presented in the section 2. We discuss the main issues in the next subsections. 87 4.1 Description Logic Ontological Engagement The first aspect to remark when embarking in such a discussion relates to the conceptual and epistemological commitment (BRACHMAN 1978) of DL as a representation formalism. Its representation purposes and consequent grounding on precise and unambiguous set theoretic semantics (BAADER ET AL, 2003) allows the language mainly to describe quite precisely domains that involve sets (as concepts or classes according to the formalism’s jargon), elements (as class and relation instances), their relations (also called properties or roles) and simple constraints that can be expressed in axioms that use a limited set of constructs like relations’ cardinality, type checking of relations’ domain and range, classes’ subsumption, complement, disjointness, equivalence, instance membership to a class and instance equality/inequality.. This is enough to make for the main reasoning task of concept classification, which means to entail class subsumption, when this relation is not declared explicitly 1. Class subsumption, equivalence, disjointness and inconsistency (standing for a class that according to its definition cannot bear any instances) can be inferred by the reasoned, also called classifier. In the next subsection we discuss the use of numerical constraints in DL. 4.2 Numerical Representations and Constraints Despite consisting of a powerful family of semantically well-founded formalisms’, DLs are not endowed with the basic mathematical resources necessary for representing neural networks, for instance the arithmetical operations. The only type of arithmetical processing is counting and comparing the number of instances that participate in relations for solving queries, like with the following axiom: Worried-Woman ≡ Woman ⊓ (≥ 3 child.Man) (6) It states that a worried woman is one who has three or more male children. The classifier is able to solve queries that searches for the instances of the class Worried- Woman by counting the instances of Woman which are related to instances of the class Man via the relation child. Although it is possible to define relations with numerical values as datatype properties in OWL (types integer and float and special types date, time and dateTime), we did not find in the literature DL extensions that address the general problem of solving arithmetical constraints, ranging from the simplest ones (like the ones that use the four basic arithmetical operations) to the most complex (using exponential, differential, integral operators, differential equations, etc). Description Logic consists indeed in a smart purely declarative formalism that performs optimized classical reasoning. Sticking to this stance, no ultimate solution is expected to the posed problem of including arbitrary arithmetical constraints, because classifier will not cope with them, ought to the fact that elementary number theory has been proved undecidable by Gödel in his classical incompleteness theorem (GÖDEL, 1 Classification based on subsumption is also used for solving equivalence and disjointness between classes and class inconsistency. Even in assertional queries, i.e., the ones which include instances, subsumption checking is the basic processing behind the reasoning (BAADER ET AL, 2003). 88 1931) as well as rational number theory, since the elementary number theory could be defined in terms of the rationals, as proved by Julia Robinson (ROBINSON, 1949). So, the hopes for proposing a general DL solution with arithmetical constraints probably lies both on restricting the constructs for a limited set of operations and solving them with built-in constructs that should be evaluated before classification (in the flavor of Prolog, for instance). Such extension, on the other hand, would change the formalism into a more procedural and less declarative one, and is not yet reported. 4.3 The Need for a Mathematical Background Knowledge in a Top Ontology Due to the low expressiveness of OWL, any top-ontology developed using it would face a lot of problems in representing complex abstract mathematical concepts. This statement is corroborated by the fact that, despite the existence of several efforts for the translation the top ontology Supper Upper-Merged Ontology (SUMO) to OWL, none of them proved successful. Translations can be found in the SUMO site but they are in OWL Full (which is known to be undecidable (BAADER ET AL, 2003)). Besides this fact, any translations of full first order logic to DL will just prune the knowledge from the original ontology that cannot be expressed in DL. We indeed have tried to work with these translations, encountering many practical problems. The two more relevant for us were the loss of a the rich set of mathematical axioms provided by SUMO upper and the necessary mathematical operations and constraints that could not be defined in DL, due to the lack of expressiveness. 5 Solution: More Expressive Languages and Top Ontologies Since DL does not fulfill our expressivity needs, the solution lies on relying on employing a more expressive language, with first-order logic (FOL) expressivity at least and, if possible, capable of defining basic arithmetic operations, and a top ontology that includes some basic mathematical definitions. A Top Ontology defines a set of generic concepts to be shared by various domain-specific ontologies. Complying with it assures semantic interoperability among the ontologies and better ontological engagement (BRACHMAN, 1978). Thus, based on the requirements enlisted above, we decided to use the SUO-KIF (Standard Upper Ontology Knowledge Interchange Format) with SUMO as the background mathematical knowledge required. The SUO-KIF language has declarative semantics and is quite comprehensive in terms of logical expressiveness, once it was intended primarily to implement the first- order logic. It is an extension of KIF (GENESERETH, 1991), created with the objective of supporting the development of SUMO. The goal was simplifying KIF, by maintaining its syntax, logical operators and quantifiers, but leaving to the ontology itself the declarations defining classes and instances, and thus eliminating the dependency from the Frame-Ontology of KIF, what happens to Eng-Math (GRUBER & OLSEN, 1994). Instead, one must create instances based on the concepts defined in SUMO as an external resource. Another relevant capability of SUO-KIF when compared to KIF’s Eng-Math is the deployment of the Sigma reasoner (PEASE, 2003). Up to now, no reasoner can take KIF’s rich expressivity on. As for the top ontology, we 89 opted for SUMO, since it endows us with the math we needed. In the next section, we present the MLP ANN ontology, with some examples using SUO-KIF and SUMO. 2. The Resulting MLP ANN Ontology in SUO-KIF In this section, we present the most important concepts of the Multi-Layer Perceptron Artificial Neural Network (MLP ANN) Ontology. We deploy the basic taxonomy of the ontology in Figure 2. Note that it uses plenty of concepts from SUMO, like the classes Abstract, Physical, Relation, BinaryRelation, IrreflexiveRelation, etc. Next, we take a deeper look at the ontology, describing some of the concepts and axioms that required a higher level of expressiveness. Figure 2. The ANN MLP Ontology basic taxonomy. 5.1 Activation function The representation of the activation function constitutes a good example of a function definition that takes advantage of the SUO-KIF expressiveness and the richness of the mathematical vocabulary provided by SUMO. In the code below, we state the concept of ActivationFunction as a function (irreflexive, intransitive, assymetric relation) whose domain is the class Neuron. (instance activationFunction BinaryPredicate) (instance activationFunction IrreflexiveRelation) (domain activationFunction 1 Neuron) (domain activationFunction 2 UnaryFunction) Then we represent the concept of a Logistic function, which is an unary function that receives a real number as the only argument and returns an MLPRealNumber . (instance Logistic UnaryFunction) (instance Logistic TotalValuedRelation) (domain Logistic 1 RealNumber) (range Logistic 1 RealNumber) 90 MLPRealNumber is defined as a real number ranging between -1 and 1: (=>(instance ?NUMBER MLPRealNumber) (and (or (greaterThan ?NUMBER -1) (equal ?NUMBER -1)) (or (lessThan ?NUMBER 1) (equal ?NUMBER 1)))) Finally we define the logistic function, as defined in subsection 2.1., def. (2): (=>(equal (Logistic ?NUMBER1) ?NUMBER2) (equal ?NUMBER2 (DivisionFn 1 (AdditionFn 1 (ExponentiationFn NumberE (SubtractionFn 0 ?NUMBER1)))))) Next, we describe the ANN learning process. 5.2 Backpropagation algorithm process The process is an iteration alternating the backward and forward phases, until the halting criterion is met. We can represent it using four axioms. The first one states that the backpropagation algorithms starts defining its net structure We can represent this process using four axioms. The first one states that for all backpropagation algorithms there exists a sub-process (the definition of the net structure) starting together with the backpropagation algorithm: (=>(and(instance ?B Backpropagation) (instance ?NSD NetStructDefinition) (subProcess ?NSD ?B)) (exists (?BL) (and(instance ?BL BackpropLoop) (subProcess ?BL ?B) (exactlyEarlier (WhenFN ?NSD) (WhenFn ?BL))))) The second axiom states that right after the net structure definition, we start the iteration over the backpropagation loop. The exactlyEarlier relation means that the first argument finishes at the same instant that the second starts. (=>(and(instance ?B Backpropagation) (instance ?BL Backproploop) (subProcess ?BL ?B) (not (StopCriteriaFn ?B))) (exists (?BL) (and(instance ?BL2 BackpropLoop) (subProcess ?BL2 ?B) (exactlyEarlier (WhenFN ?BL)(WhenFn?BL2))))) Finally, the last axiom states that, if the halting criterion is met, the algorithm ends. The finishes relation means that both arguments finish at the same time: (=>(and(instance ?B Backpropagation) (instance ?BL BackpropLoop) (subProcess ?BL ?B) (StopCriteriaFn ?B)) (finishes (WhenFn ?B) (WhenFn ?BL))) 91 The weights are updated (subsection 2.2., definition 5) according to the axiom shown below. It states that if a weight is identified as one to be updated, then the old weight holds only until the loop cycle lasts, and the new weight is calculated and set. (=> (and (instance ?UPDATEWEIGHT UpdateSynapseWeight) (synapseToUpdate ?UPDATEWEIGHT ?SYNAPSE) (connects ?SYNAPSE ?NODEA ?NODEB) (hasOutput ?NODEB ?OUTPUT) (holdsDuring (BeginFn ?UPDATEWEIGHT) (hasWeight ?SYNAPSE ?OLDWEIGHT))) (holdsDuring (EndFn ?UPDATEWEIGHT) (and (hasWeight ?SYNAPSE ?NEWWEIGHT) (equal ?NEWWEIGHT (AdditionFn ?OLDWEIGHT (MultiplicationFn ?LEARNRATE (MultiplicationFn ?OUTPUT (NeuronErrorFn ?SYNAPSE)))))))) 5.3 An Example of Ternary Relation: The Synapse Definition In the next encodings, we show an example of a ternary relation, the relation connects, that links two neurons through a synapse. The first stretch of code shows the basic characterizations of the relation in terms of inputs and outputs: (instance connects TernaryPredicate) (domain connects 1 Synapse) (domain connects 2 Neuron) (domain connects 3 Neuron) Next, an axiom is defined, stating that all connections of a neuron are linked with a neuron from its next layer. Note that a synapse takes part of the ternary relation, thus sufficing for our representation needs. (=>(and(instance ?N-A Neuron) (connects ?SYNAPSE ?N-A ?N-B)) (exists(?LAYERA ?LAYERB) (and(instance ?LAYERB Layer) (instance ?LAYERA Layer) (neuronLayer ?N-A ?LAYERA) (neuronLayer ?N-B ?LAYERB) (nextLayer ?LAYERA ?LAYERB)))) In the next section, we discuss related work regarding mathematical ontologies. 92 5.4 Discussion The ontology defines the exact constraints that hold among the many elements of a MLP ANN, as well as the calculations and sequential operations needed to update it during the training phase. Once SUO-KIF is endowed with a reasoned able to deal with the mathematical constraints, an application that employs it – an intelligent agent, expert system, computer algebra system or theorem prover - is capable of creating a virtual MLP ANN, run it (including all the calculations and updates), and - more important in terms of declarativity – answer queries about each aspect of it, like stating how many layers at least a MLP ANN should possess, etc. With that features, even an intelligent tutor system can make a good use from its knowledge. Note that, when an expressive formalism is used with a supportive top ontology, most, if not all, mathematical knowledge can be represented by a similar solution. Simple examples are other types of neural networks (even recurrent, constructive) 6 Related Work In the field of Artificial Intelligence, the first successful systems to deal with mathematical contents that go beyond simple numerical calculi came from the branch known as Computer Algebra Systems (CAS) (BERTOLI ET AL 1998). They are capable of performing algebraic and symbolic computation in an abstract way A representative example of this trend was the REDUCE system (HEARN 2004), which is still in use nowadays. It is able to calculating algebraic derivates and integrals of complex functions among many other functionalities. Many of those systems employ declarative solutions for algebraic problems and some of them are able of using and presenting proofs indeed. A detailed comparison among computer algebra systems and their features can be found in [13]. Nevertheless, the integration between this type of system and the growing field of ontology is still lacking tough. The consequence is that some valuable mathematical knowledge available in mathematical ontologies, like the characterization of distinct relation types used here, are not being fully exploited in CAS systems. As for the problem of mathematical knowledge in ontologies, we have searched the literature for possible solutions to our problems, before deciding for the SUO-KIF language. We found solutions ranging from simple syntactic agreements, like MathML (CARLISLE 2009) - a markup language without axioms -, to full-fledged languages and top ontologies like SUO-KIF and SUMO. Among them, we found a proposal to mix OWL with OpenMath to encode math contents (BUSWELL 2004). We indeed made attempts to employ this approach, by to representing the math needed in the MLP ANN ontology in OWL and processes using the Semantic Web Rule language (SWRL) built- ins. However, none of these solutions fitted our needs. All of them were limited, in the sense of depending upon external features (like the rules in SWRL), instead of using a mathematical theory. They mostly represent the expressions in a structured way and even the solutions using rules will produce fragmented pieces of knowledge that could pose problems of maintenance in the future. Furthermore, we tried out a more consistent approach, the use of EngMath (GRUBER & OLSEN, 1994), a mathematical ontology for engineering. EngMath could indeed constitute a sound alternative, as it takes advantage of the Kif-Numbers ontology 93 (GRUBER & OLSEN, 1994), a KIF vocabulary focused on numbers, arithmetic operations and related definitions. We chose SUMO and SUO-KIF because, we conclude that SUMO covers all definitions comprised in Kif-Numbers and EngMath. 7 Future Work and Conclusions Despite the popularity and usefulness of DL languages, we claim that this formalism suffers from a lack of expressiveness for the representation of mathematical knowledge. In our practical experience portrayed here, the most relevant lesson learned was that we could only overcome DL representational problems by using at least a full first-order logic formalism with a supportive top ontology like SUMO, that contains the mathematical background knowledge required to qualify and define properly the math definitions for the new ontologies. We presented a thorough case study in that direction, on the field of automatic learning, the MLP ANN ontology. We envisage two types of future work. As for the use of the developed ontology, we are heading for practical applications of the ontology. The creation of these applications, such as an intelligent agent capable of interacting with users answering questions about artificial neural networks (making use of the ontology) are in our agenda. We also consider the translation of the ontology to other languages, such as Prolog, so as to enable us to reason with it and run concrete applications. Another more general future work lies on the investigation of ways to embed ontologies into computer algebra systems. Devising a solution for this general problem will certainly endow the latter with more powerful reasoning techniques while solving mathematical problems. For instance, CAS systems could take advantage during inference of the applicable mathematical constraints defined as relations qualifiers in the ontology. These constraints need not be hardcoded in the systems, thus increasing knowledge reuse. Acknowledgments. The authors would like to thank prof. Richard Banach, from the University of Manchester, England, who provided us with valuable input on the limitations of Description Logic for Mathematical representations, and prof. Jacques Calmet, from the University of Karlsruhe, Germany, who, as scientific ancestor, indirectly introduced us to the field of Computer Algebra. References BAADER, F., CALVANESE, D., MCGUINNESS, D.L., NARDI, D., PATEL- SCHNEIDER, P.F., eds.: The Description Logic Handbook. Cambridge Univ. Press (2003) BERTOLI, P.G., CALMET, J., GIUNCHIGLIA, F., HOMANN, K.: Specification and integration of theorem provers and computer algebra systems. In Calmet, J., Plaza, J., eds.: AI and Symbolic Computation: Proceedings of AISC'98, Berlin, Germany, Springer (1998) 94-106 BRACHMAN, R.J.: On the epistemological status of semantic networks. BBN Report 3807, Bolt Beranek and Newman Inc., (apr 1978) BUSWELL, S. CAPROTTI, O.,CARLISLE, D.,.DEWAR D., GAËTANO, M., KOHLHASE, M.: Open math standard report, Open Math Society (2004) 94 CARLISLE, D., MINER, R., ION, P.: Mathematical markup language (MathML) version 3.0. Candidate recommendation, W3C (dec 2009) GENESERETH, M.R.: Knowledge interchange format. In Allen, J.F., Fikes, R., Sandewall, E., eds.: Principles of Knowledge Representation and Reasoning. Morgan Kaufmann, San Mateo, California (1991) 599-600 GÖDEL, K.: Über formal unentscheidbare sätze der principia mathematica und verwandter systeme. Monatshefte für Mathematik und Physik 38 (1931) 173-198 GRUBER, T.R., OLSEN, G.R.: An ontology for engineering mathematics. In Doyle, J., Sandewall, E., Torasso, P., eds.: Principles of Knowledge Representation and Reasoning. Morgan Kaufmann, San Francisco HAYKIN, S.: Neural networks. MacMillan, New York (1994) HEARN, A.C.: Reduce user's and contributed packages manual, version 3.8. Report, The RAND Corporation, Berlin, Germany (2004) NILES, I., PEASE, A.: Towards a standard upper ontology. In: Proc. of the int.conference on Formal Ontology in Information Systems, New York, NY, USA, ACM (2001) 2-9 PEASE, A. The Sigma Ontology Development Environment. In Working Notes of the IJCAI-2003 Workshop on Ontology and Distributed Systems, Acapulco, Mexico, 2003 PEASE,A. A Standard Upper Ontology Knowledge Interchange Format. 2009 ROBINSON, J.: Definability and decision problems in arithmetic. Journal of Symbolic Logic 14 (1949) 98-114 WELTY, C., MCGUINNESS, D.L., SMITH, M.K.: OWL web ontology language guide. W3C recommendation, W3C (2004) http://www.w3.org/TR/2004/RECowl- guide-20040210/. WIKIPEDIA: Comparison of computer algebra systems. en.wikipedia.org/wiki/Comparison of computer algebra systems, 2010. 95