<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>New approaches to ATPs</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Frieder Simon</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Oxford, UK National Institute of Informatics</institution>
          ,
          <country country="JP">Japan</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>carried out as part of my internship at the National Institute of Informatics</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>In this paper a high-level outline of two research directions is presented, aimed at improving current, automated proof systems. When it comes do doing mathematics-as opposed to communicating mathematics, for which LATEX, citeable article databases and other instruments have become the standard-there do not exist general-purpose tools: Various symbolic algebra programs, numerical analysis suites or other more specialized tools are employed, but these are typically heavily domain-dependent. Compared to these, at the other end of the spectrum there are automated theorem provers (ATPs), which work in full generality, and could be considered to be the ultimate tool to a working mathematician; though currently they are far from being able to attain human level proving performance. Research in ATP can roughly be divided in two approaches: One works by encoding mathematics in a formal language and reasoning on a formal level. Recently, machine learning has been successfully used in this regard to augment existing ATP approaches (e.g., lemma selection) by improving the proof search mechanism through learning; see, e.g., [7, 6] for reinforcement learning or [2] deep learning approaches. The other approach works by completely discarding modeling mathematics with a purely formal language, and building provers whose focus is not on soundness, but on proving more advanced statements, accepting a non-zero probability that those proofs might be wrong. Here we mention [5] as a modern candidate. (We will discuss why relaxing the soundness requirement might be meaningful at the end of the next section.) In this text we will outline how advancements in either direction might be made.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Examples of such are, e.g., [
        <xref ref-type="bibr" rid="ref9">10</xref>
        ] or proof libraries associated to interactive theorem provers, such as fragments
from Isabelle’s Archive of Formal Proofs (http://afp.sf.org), in addition to the existing libraries specifically
curated for reinforcement learning, that come with the mentioned papers above.
      </p>
      <p>
        Regarding the second approach we propose to build a productivity tool, a mathematical knowledge miner, that
could be integrated in an ATP of the second kind, similar to [
        <xref ref-type="bibr" rid="ref4">5</xref>
        ]. The goal of the mathematical knowledge miner
is to map the content of a piece of stand-alone mathematical text, such as a research article, onto a knowledge
representation formalism, such as an ontology or a knowledge graph, and to thereby provide an understanding of
the main theorem of the text from context it is embedded in. The starting point of learning such a mapping could
be an approach as in [
        <xref ref-type="bibr" rid="ref8">9</xref>
        ] (combined with [1]). Different avenues in related fields of natural language processing
have to be explored in order to find the optimal approach; this will make heavy use of modern deep learning
methods from the field of natural language processing.
      </p>
      <p>
        In case of knowledge graphs, every node could be either a definition, a theorem/lemma/proposition or an
example and arrows should indicate how these are related to each other. Because this is an ambitious undertaking,
it would be appropriate to first focus on an easy case, where the article in question is not a modern research
article, but an article as they can be found in the American Mathematical Monthly [
        <xref ref-type="bibr" rid="ref2">3</xref>
        ] magazine, which caters
to a wide, mathematically literate audience.
      </p>
      <p>
        The novelty of the proposal does not lie in creating such knowledge graphs or ontologies per se, as similar
solutions that classify the contents of a research in terms already exist, see, e.g., [
        <xref ref-type="bibr" rid="ref7">8</xref>
        ]. Rather it lies in using them
in the context of an ATP. Such a “compressed view” of a mathematical text, captured by a knowledge graph or
an ontology, would cover a new, middle ground between the formulation of the contents of the article in natural
language and the formulation of the contents in a finely grained formal language, as would be necessary if one
would want current ATPs to parse the article.
      </p>
      <p>
        It is conceivable, in order to build an ATP of the second kind, that a succession of tools that provide different
higher level features of a mathematical document should first be devised, as exemplified by a knowledge graph
or ontology tailored to ATP needs. Building modules that enable deep neural networks to be more successful
at understanding numeracy, a starting point being [
        <xref ref-type="bibr" rid="ref10">11</xref>
        ], would be a further possibility for a bottom-up approach
to work towards an ATP. Once such a succession of tools leading up to a prover would be in place, it seems
reasonable to assume that one could do high-level reasoning in such an ontology of mathematical objects and
infer high-level information. One—perhaps controversial—aspect of this approach will be to not require such a
prover to be sound. This might be unusual, but it is an approach that is worth exploring: Requiring soundness
is one of the reasons current ATPs are designed in a purely formal language, thereby producing only completely
rigorous theorems (unless the Trusted Computing Module an ATP uses to do the inference has bugs). But
experience shows that this approach leads to ATPs that have difficulty proving formalized versions of advanced
theorems (as they can be found, for example, in graduate textbooks). Another reason not to insist on soundness
is that mathematicians are also not concerned by being presented with a proof that might not be sound, since
they can check soundness for themselves. Rather, it is a matter of much greater importance to allow the ATP
to come up with any proof at all, that can be easily read (and checked) by a human.
3
      </p>
    </sec>
    <sec id="sec-2">
      <title>Summary</title>
      <p>We conclude this vision paper by underlining that machine learning is the prevalent method that will be used
to devise new approaches to advance ATPs. On one hand this could be accomplished by improving current
ATP systems that work on a fine-grained formal language, encoding mathematics. On the other hand machine
learning approaches could help pave the way towards non-formal ATPs.</p>
      <p>Being still at the beginning of my doctoral research, it is clear that realizing the described approaches is
an ambitious undertaking, given the difficulty of the domain of automated theorem proving. However, these
approaches have not yet been fully explored, so there is reason to be optimistic that at least one of the approaches
will be fruitful.
[1] L. Abzianidze, LANGPRO. Natural Language Theorem Prover, Proceedings of the 2017 EMNLP System</p>
      <p>Demonstrations, pp 115–120</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Alemi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Chollet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Een</surname>
          </string-name>
          , G. Irving,
          <string-name>
            <given-names>C.</given-names>
            <surname>Szegedy</surname>
          </string-name>
          , J. Urban, DeepMath
          <article-title>- Deep Sequence Models for Premise Selection</article-title>
          , Advances in
          <string-name>
            <surname>NIPS</surname>
          </string-name>
          , Vol.
          <volume>29</volume>
          ,
          <year>2016</year>
          , pp.
          <fpage>2235</fpage>
          -
          <lpage>2243</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>American</given-names>
            <surname>Mathematical</surname>
          </string-name>
          <string-name>
            <surname>Monthly</surname>
          </string-name>
          , https://www.maa.org/press/ ...
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>J.A.</given-names>
            <surname>Arjona-Medina</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Gillhofer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Widrich</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Unterthiner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Brandstetter</surname>
          </string-name>
          , S. Hochreiter, RUDDER. Return Decomposition of Delayed Rewards, https://arxiv.org/abs/
          <year>1806</year>
          .07857
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>M.</given-names>
            <surname>Ganesalingam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Gowers</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A Fully</given-names>
            <surname>Automatic</surname>
          </string-name>
          <article-title>Theorem Prover with Human-Style Output</article-title>
          ,
          <source>Journal of Automated Reasoning February</source>
          , Vol.
          <volume>58</volume>
          (
          <issue>2</issue>
          ),
          <year>2017</year>
          , pp
          <fpage>253</fpage>
          -
          <lpage>291</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>D.</given-names>
            <surname>Huan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Dhariwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <surname>I.</surname>
          </string-name>
          <article-title>Sutskever, GamePAD. A Learning Environment for Theorem proving</article-title>
          ,
          <source>ICLR</source>
          <year>2019</year>
          , https://openreview.net/forum?id=r1xwKoR9Y7
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>C.</given-names>
            <surname>Kaliszyk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Urban</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Michalewski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Olšák</surname>
          </string-name>
          ,
          <article-title>Reinforcement Learning of Theorem Proving</article-title>
          ,
          <source>Proceedings of the 32nd International Conference on NIPS</source>
          ,
          <year>2018</year>
          , pp
          <fpage>8836</fpage>
          -
          <lpage>8847</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>M.</given-names>
            <surname>Liakata</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Saha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Dobnik</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Batchelor</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Rebholz-Schuhmann</surname>
          </string-name>
          ,
          <article-title>Automatic recognition of conceptualization zones in scientific articles and two life science applications</article-title>
          ,
          <source>Bioinformatics</source>
          , Vol.
          <volume>28</volume>
          (
          <issue>7</issue>
          ),
          <year>2012</year>
          , pp
          <fpage>991</fpage>
          -
          <lpage>1000</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.-W.</given-names>
            <surname>Huang</surname>
          </string-name>
          &amp;
          <string-name>
            <surname>A. Courville</surname>
          </string-name>
          ,
          <article-title>Neural Language Modelling by Jointly Learning Syntax and Lexicon</article-title>
          ,
          <string-name>
            <surname>ICLR</surname>
          </string-name>
          <year>2018</year>
          , https://openreview.net/forum?id=rkgOLb-
          <fpage>0W</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>G.</given-names>
            <surname>Sutcliffe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Suttner</surname>
          </string-name>
          ,
          <source>The TPTP Problem Library for Automated Theorem Proving, Journal of Automated Reasoning</source>
          , Vol.
          <volume>21</volume>
          (
          <issue>2</issue>
          ),
          <year>1998</year>
          , pp
          <fpage>177</fpage>
          -
          <lpage>203</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>A.</given-names>
            <surname>Trask</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Hill</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Reed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Rae</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Dyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Blunsom</surname>
          </string-name>
          ,
          <source>Neural Arithmetic Logic Units, Advances in NIPS 2018</source>
          , Vol.
          <volume>31</volume>
          ,
          <year>2018</year>
          , pp.
          <fpage>8035</fpage>
          -
          <lpage>8044</lpage>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>