<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Algorithmically Generating New Algebraic Features of Polynomial Systems for Machine Learning</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Dorian Florescu Dorian.Florescu@coventry.ac.uk</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Faculty of Engineering, Environment and Computing, Coventry University</institution>
          ,
          <addr-line>Coventry, CV1 5FB</addr-line>
          ,
          <country country="UK">UK</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Matthew England</institution>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2019</year>
      </pub-date>
      <abstract>
        <p>There are a variety of choices to be made in both computer algebra systems (CASs) and satis ability modulo theory (SMT) solvers which can impact performance without a ecting mathematical correctness. Such choices are candidates for machine learning (ML) approaches, however, there are di culties in applying standard ML techniques, such as the e cient identi cation of ML features from input data which is typically a polynomial system. Our focus is selecting the variable ordering for cylindrical algebraic decomposition (CAD), an important algorithm implemented in several CASs, and now also SMT-solvers. We created a framework to describe all the previously identi ed ML features for the problem and then enumerated all options in this framework to automatically generate many more features. We validate the usefulness of these with an experiment which shows that an ML choice for CAD variable ordering is superior to those made by human created heuristics, and further improved with these additional features. This technique of feature generation could be useful for other choices related to CAD, or even choices for other algorithms in CASs / SMT-solvers with polynomial systems as input.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        In practice such choices may be made by man-made heuristics based on some experimentation (e.g. [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]) or
magic constants where crossing a single threshold changes system behaviour [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. It is likely that many of these
decisions could be improved by allowing learning algorithms to analyse the data. The broad topic of this paper
is ML for algorithm choices where the input is a set of polynomials, which encompasses a variety of tools in
computer algebra systems and the SMT theory of [Quanti er Free] Non-Linear Real Arithmetic, [QF]NRA.
      </p>
      <p>
        There has been little research on the use of ML in computer algebra: only [
        <xref ref-type="bibr" rid="ref33">33</xref>
        ] on identifying e cient
multivariate Horner schemes; [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ] [
        <xref ref-type="bibr" rid="ref27">27</xref>
        ] [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ] on the topic of CAD variable ordering choice; [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ], [
        <xref ref-type="bibr" rid="ref27">27</xref>
        ] on the question of
whether to precondition CAD with Groebner Bases; and [
        <xref ref-type="bibr" rid="ref31">31</xref>
        ] on deciding the order of sub-formulae solving for
a QE procedure. Within SMT there has been signi cant work on the Boolean logic side e.g. the portfolio SAT
solver SATZilla [
        <xref ref-type="bibr" rid="ref46">46</xref>
        ] and MapleSAT [
        <xref ref-type="bibr" rid="ref34">34</xref>
        ] which views solver branching as an optimisation problem. However
there is little work on the use of ML to choose or optimise theory solvers.
      </p>
      <p>
        We note that other elds of mathematical software are ahead in the use of ML, most notably the automated
reasoning community (see e.g. [
        <xref ref-type="bibr" rid="ref43">43</xref>
        ], [
        <xref ref-type="bibr" rid="ref32">32</xref>
        ], [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], or the brief survey in [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ]).
1.1
      </p>
      <sec id="sec-1-1">
        <title>Di culties with ML for problems in NRA</title>
        <p>
          There are di culties in applying standard ML techniques to problems in NRA. One is the lack of su ciently
large datasets, which is addressed only partially by the SMT-LIB. The experiment in [
          <xref ref-type="bibr" rid="ref26">26</xref>
          ] found the [QF]NRA
sections of the SMT-LIB too uniform, and had to resort to random generated examples (although the state
of benchmarking in computer algebra is far worse [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ]). There have been improvements since then, with the
benchmarks increasing both in number and diversity of underlying application. For example, there are now
problems arising from biology [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ], [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ] and economics [
          <xref ref-type="bibr" rid="ref38">38</xref>
          ], [
          <xref ref-type="bibr" rid="ref39">39</xref>
          ].
        </p>
        <p>Another di culty is the identi cation of suitable features from the input with which to train the ML models.
There are some obvious candidates concerning the size and degrees of polynomials, and the distribution of
variables. However, this provides a starting set (i.e. before any feature selection takes place) that is small in
comparison to other machine learning applications. The main focus of this paper is to introduce a method to
automatically (and cheaply) generate further features for ML from polynomial systems.
1.2</p>
      </sec>
      <sec id="sec-1-2">
        <title>Contribution and plan</title>
        <p>Our main contributions are the new feature generation approach described in Section 3 and the validation of its
use in the experiments described in Sections 4 5. The experiments are for the choice of variable ordering for
cylindrical algebraic decomposition, a topic whose background we rst present in Section 2, but we emphasise
that the techniques may be applicable more broadly.
2
2.1</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>Background on variable ordering for CAD</title>
      <sec id="sec-2-1">
        <title>Cylindrical algebraic decomposition</title>
        <p>A Cylindrical Algebraic Decomposition (CAD) is a decomposition of ordered Rn space into cells arranged
cylindrically : the projections of any pair of cells with respect to the variable ordering are either equal or disjoint.
The projections form an induced CAD of the lower dimensional space. The cells are (semi)-algebraic meaning
each can be described with a nite sequence of polynomial constraints.</p>
        <p>A CAD is produced to be truth-invariant for a logical formula (so the formula is either true or false on each
cell). Such a decomposition can then be used to perform Quanti er Elimination (QE) over the reals, i.e. given
a quanti ed Tarski formula nd an equivalent quanti er free formula over the reals. For example, QE would
transform 9x; ax2 + bx + c = 0 ^ a 6= 0 to the equivalent unquanti ed statement b2 4ac 0. A CAD over
the (x; a; b; c)-space could be used to ascertain this, so long as the variable ordering ensured that there was an
induced CAD of (a; b; c)-space. We test one sample point per cell and construct a quanti er free formula from
the relevant semi-algebraic cell descriptions.</p>
        <p>
          CAD was introduced by Collins in 1975 [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] and works relative to a set of polynomials. Collins' CAD produces
a decomposition so that each polynomial has constant sign on each cell (thus truth-invariant for any formula
built with those polynomials). The algorithm rst projects the polynomials into smaller and smaller dimensions;
and then uses these to lift to incrementally build decompositions of larger and larger spaces according to the
polynomials at that level. For further details on CAD see for example the collection [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ].
        </p>
        <p>
          QE has numerous applications throughout science and engineering [
          <xref ref-type="bibr" rid="ref42">42</xref>
          ]. Our work also speeds up independent
applications of CAD, such as reasoning with multi-valued functions [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ] or motion planning [
          <xref ref-type="bibr" rid="ref45">45</xref>
          ].
The de nition of cylindricity and both stages of the algorithm are relative to an ordering of the variables. For
example, given polynomials in variables ordered as xn xn 1 : : : ; x2 x1 we rst project away xn and so
on until we are left with polynomials univariate in x1. We then start lifting by decomposing the x1 axis, and
then the (x1; x2) plane and so so on. The cylindricity condition refers to projections of cells in Rn onto a space
(x1; : : : ; xm) where m &lt; n. There have been numerous advances to CAD since its inception: new projection
schemes [
          <xref ref-type="bibr" rid="ref35">35</xref>
          ], [
          <xref ref-type="bibr" rid="ref37">37</xref>
          ]; partial construction [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ], [
          <xref ref-type="bibr" rid="ref44">44</xref>
          ]; symbolic-numeric lifting [
          <xref ref-type="bibr" rid="ref41">41</xref>
          ], [
          <xref ref-type="bibr" rid="ref29">29</xref>
          ]; adapting to the Boolean
structure [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ], [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ]; and adaptations for SMT [
          <xref ref-type="bibr" rid="ref30">30</xref>
          ], [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]. However, the need for a xed variable ordering remains.
        </p>
        <p>Depending on the application, the variable ordering may be determined, constrained, or free. QE, requires
that quanti ed variables are eliminated rst and that variables are eliminated in the order in which they are
quanti ed so there is only partial freedom. The discriminant in the example from Section 2.1 could have been
found with any CAD which eliminates x rst. A CAD for the quadratic polynomial under ordering a b c
has only 27 cells, but 115 are required for the reverse ordering.</p>
        <p>Since we can switch the order of quanti ed variables in a statement when the quanti er is the same, we also
have some choice on the ordering of quanti ed variables. For example, a QE problem of the form 9x9y8a (x; y; a)
could be solved by a CAD under either ordering x y a or ordering y x a. In the SMT context there is
only a single existentially quanti ed block and so there is a completely free choice of ordering.</p>
        <p>
          The choice of variable ordering can have a great e ect on the time / memory use of CAD, and the number
of cells in the output. Brown and Davenport presented a class of problems in which one variable ordering gave
output of double exponential complexity in the number of variables and another output of a constant size [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ].
2.3
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>Prior work on choosing the variable ordering</title>
        <p>
          Heuristics have been developed to choose a variable ordering, with Dolzmann et al. [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ] giving the best known
study. After analysing a variety of metrics they proposed a heuristic, sotd, which constructs the full set of
projection polynomials for each permitted ordering and selects the ordering whose corresponding set has the lowest
sum of total degrees for each of the monomials in each of the polynomials. The second author demonstrated
examples for which that heuristic could be misled in [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ]; and then later showed that tailoring to an
implementation could improve performance [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ]. These heuristics all involved potentially costly projection operations on
the input polynomials.
        </p>
        <p>
          In [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ] the second author of the present paper collaborated to use a support vector machine to choose which
of three human made heuristics to believe when picking the variable ordering, based only on simple features of
the input polynomials. The experiments identi ed substantial subclasses on which each of the three heuristics
made the best decision, and demonstrated that the machine learned choice did signi cantly better than any one
heuristic overall. This work was picked up again in [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ] by the present authors, where ML was used to predict
directly the variable ordering for CAD, leading to the shortest computing time, with experiments conducted for
four di erent ML models.
        </p>
        <p>
          Both [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ] and [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ] used a set of 11 human identi ed features. These did lead to good performance of the
models, with ML outperforming the prior human created heuristics, but a starting set of 11 features is relatively
small for ML and so we hypothesise that identifying more would improve the results.
3
3.1
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Generating new features algorithmically</title>
      <sec id="sec-3-1">
        <title>Existing features for sets of polynomials</title>
        <p>
          An early heuristic for the choice of CAD variable ordering is that of Brown [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ], which chooses a variable ordering
according to the following criteria, starting with the rst and breaking ties with successive ones.
(1) Eliminate a variable rst if it appears with the lowest overall degree in the input.
(2) For each variable calculate the maximum total degree for the set of terms in the input in which it occurs.
        </p>
        <p>
          Eliminate rst the variable for which this is lowest.
(3) Eliminate a variable rst if there is a smaller number of terms in the input which contain the variable.
Despite being computationally cheaper than the sotd heuristic (because the latter performs projections before
measuring degrees) experiments in [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ] suggested this simpler measure actually performs slightly better, although
mmmaaaxxxmmm;;;ppp ddd21mm;;pp
        </p>
        <p>3
avp (sgn (Pm d1m;p))
avp (sgn (Pm d2m;p))
avp (sgn (P 3m;p))
avm;p (sgn (dm1md;p))
avm;p (sgn (d2m;p))
avm;p (sgn (d3m;p))
the key message from those experiments is that there were substantial subsets of problems for which each
humanmade heuristic made a better choice than the others.</p>
        <p>
          The Brown heuristic inspired almost all the features used by the authors of [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ], [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ] to perform ML for CAD
variable ordering, with the full set of 11 features listed in Table 1 (column 3 will be explained later).
3.2
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>A new framework for generating polynomial features</title>
        <p>
          Our new feature generation procedure is based on the observation that all the measurements taken by the Brown
heauristic, and all those features used in [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ], [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ] can be formalised mathematically using a small number of
functions. For simplicity, the following discussion will be restricted to polynomials of 3 variables as these were
used in the following experiments, but everything generalises in an obvious way to n variables.
        </p>
        <p>Let a problem instance P r be de ned by a set of P polynomials</p>
        <p>P r = fPp j p = 1; : : : ; P g:
This is the typical input for producing a sign-invariant CAD. Of course, any problem instance consisting of a
logical formula whose atoms are polynomial sign conditions can also have such a set extracted.</p>
        <p>In the following we de ne the notation for polynomial variables and coe cients that will be used throughout
the manuscript. Each polynomial with index p, for p = 1; : : : ; P contains a di erent number of monomials, which
will be labelled with index m, where m = 1; : : : ; Mp and Mp denotes the number of monomials in polynomial
p. We note that these are just labels and are not setting an ordering themselves. The degrees corresponding to
each of the variables x1; x2; x3 are a function of m and p. These need to be explicitly labelled in order to allow
a rigorous de nition of our proposed procedure of feature generation.</p>
        <p>We can now write each polynomial as</p>
        <p>Pp =
m=1
Mp dm;p dm;p dm;p</p>
        <p>X cm;p x11 x22 x33 ; p = 1; : : : ; P:
Here, xv represents the polynomial variables (v = 1; 2; 3). Thus for each monomial in each polynomial there is a
tuple (m; p) of positive integers that label it. Then in turn we denote by dm;p the degree of variable xv in that
v
monomial, and by cm;p the constant coe cient, i.e., tuple superscripts are giving a label for a monomial in a
problem instance. The original indices are simply a labelling and not an ordering of the variables x1; x2; x3:
Therefore, any one of our problem instances P r is uniquely represented by a set of sets</p>
        <p>SP r =</p>
        <p>[cm;p; (d1m;p; d2m;p; d3m;p)] j m = 1; : : : ; Mp j p = 1; : : : ; P :</p>
        <p>Observe that each of Brown's measures can now be formalised as a vector of features for choosing a variable.
(1) Overall degree in the input of a variable: maxm;p dvm;p.
(1)
(2)
(3)
(2) Maximum total degree of those terms in the input in which a variable occurs:</p>
        <p>maxm;p sgn(dvm;p) (d1m;p + d2m;p + d3m;p).
(3) Number of terms in the input which contain the variable: Pm;p sgn(dvm;p).</p>
        <p>In the latter two we use the sign function to discriminate between monomials which contain a variable (sign of
degree is positive) and those which do not (sign of degree is zero). Of course the sign of the degree is never
negative.</p>
        <p>De ne now also the averaging functions
avm ,
1</p>
        <p>X;
Mp m
avp , P1 Xp;
avm;p ,
1 X
P p
1</p>
        <p>X :
Mp m
Then all the features in Table 1 can be formalised similarly to Brown's metrics, as shown in the third column of
Table 1. We can place all of these scalars into a single framework:
f (P r) = (g4 g3 g2 g1
hm;p) (P r) ;
(4)
where
and g1; g2; g3, and g4 are all taken from the set
hm;p (P r) 2</p>
        <p>dvm;p; sgn (dvm;p) (Pv0 dvm0;p) j v = 1; 2; 3
maxp; maxm; maxm;p; max0; Pp; Pm; Pm;p; P0; avp; avm; avm;p; av0; sgn; sgn0 :
In the above set max0, P0, av0 and sgn0 are all equal to the identity function.</p>
        <p>For example, let P r = fx12x2 x3; x1x24x32 + x1x3g. If m = 1; p = 2, then h1;2 (P r) 2
Pv0 d1v;02 j v = 1; 2; 3 = 1; 1 7; 4; 4 7; 2; 2 7 .
d1v;2; sgn d1;2
v
3.3</p>
      </sec>
      <sec id="sec-3-3">
        <title>Generating additional features</title>
        <p>We will thus consider deriving all of the other features which fall into this framework, but to do so we must rst
impose a number of rules.</p>
        <p>1. The functions g1; g2; g3; g4 must all belong to distinct categories of function, i.e. one each of max, P, av,
and sgn.
2. Exactly one of the functions g1; g2; g3; g4 is computed over p and exactly one is computed over m (it may
be the same one).
3. The computation over p is always performed by a function gi with an index i greater or equal to that of the
function computing over m.</p>
        <p>The expression of f (P r) can be interpreted as follows. The values hm;p (P r) are functions of variables m
and p. Each of the functions g1; g2; g3; g4 either leave the function unchanged, or they turn it into a function of
fewer variables ( rst into a function of p, and then into a scalar value, representing the ML feature).</p>
        <p>The rules above are justi ed as follows. Rule 1 reduces the redundancy in the feature set. Rules 2 and 3
guarantee that the feature fv (P r) is well de ned and is a scalar number. In particular, Rule 3 is necessary
because the computation over the terms in a polynomial is dependent on their number, which is not the same
for all polynomials.</p>
        <p>The nal set ff (1)(P r); : : : ; f (Nf )(P r)g has size Nf = 1728 for a problem with 3 variables. This number is
attained as follows: we have 12 possible distributions of indexes to the functions g1; : : : ; g4 as shown in Table 2;
then 4! possible orderings of those functions; and 6 possible choices for h. So 4! 6 12 = 1728.</p>
        <p>However, many of these features will be identical (e.g. due to a di erent placement of the identify function).
We do not identify these manually now: that task is trivial for a given dataset, but substantial to do in generality.
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Machine learning experiment with the new features</title>
      <p>
        We now describe a ML experiment to choose the variable ordering for cylcindrical algebraic decomposition. The
methodology here is similar to that in our recent paper [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ] except for the addition of the extra features from
Section 3. A more detailed discussion of the methodology can be found in [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ].
We use the nlsat dataset1 produced to evaluate the work in [
        <xref ref-type="bibr" rid="ref30">30</xref>
        ], thus the problems are all fully existentially
quanti ed. Although there are CAD algorithms that reduce what is being computed based on the quanti ers in
the input (most notably via Partial CAD [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]), the conclusions drawn are likely to be applicable outside of the
SAT context. We use the 6117 problems with 3 variables from this database, so each has a choice of six di erent
variable orderings. We extracted only the polynomials involved, and randomly divided into two datasets for
training (4612) and testing (1505). Only the former was used to tune the parameters of the ML models.
We used the CAD routine CylindricalAlgebraicDecompose which is part of the RegularChains Library for
Maple. This algorithm builds decompositions rst of n-dimensional complex space before re ning to a CAD of
Rn [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. We ran the code in Maple 2018 but used an updated version of the RegularChains Library
(http://www.regularchains.org). Training and evaluation of the ML models was done using the scikit-learn
package [
        <xref ref-type="bibr" rid="ref40">40</xref>
        ] v0.20.2 for Python 2.7. The features for ML were extracted using code written in the sympy package
v1.3 for Python 2.7, as was the Brown heuristic. The sotd heuristic was implemented in Maple as part of the
ProjectionCAD package [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ].
      </p>
      <p>CAD construction was timed in a Maple script that was called separately from Python for each CAD (to avoid
Maple's caching of results). The target variable ordering for ML was de ned as the one that minimises the
computing time for a given problem. All CAD function calls included a time limit. For the training dataset an
initial time limit of 4 seconds was used, doubled incrementally if all orderings timed out, until CAD completed
for at least one ordering (a target variable ordering could be assigned for all problems using time limits no bigger
than 64 seconds). The problems in the testing dataset were processed with a larger time limit of 128 seconds for
all orderings (timeout set as 128s).
4.4</p>
      <sec id="sec-4-1">
        <title>Feature simpli cation</title>
        <p>
          When computed on a set of problems fP r1; : : : ; P rN g, some of the features f (i) turn out to be constant, i.e.
f (i)(P r1) = f (i)(P r2) = = f (i)(P rN ): Such features will have no bene t for ML and are removed. Further,
other features may be repetitive, i.e. f (i)(P rn) = f (j)(P rn); 8n = 1; : : : ; N: This repetition may represent a
mathematical equality, or just be the case of the given dataset. Either way, they are merged into a single feature
for the experiment. After these steps, we are left with 78 features for our dataset: so while a large majority were
redundant, we still have seven times those available in [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ], [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ].
4.5
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>Feature selection</title>
        <p>
          Feature selection was performed with the training dataset to see if any features were redundant for the ML. We
chose the Analysis of Variance (ANOVA) F-value to determine the importance of each feature for the classi cation
task. Other choices we considered were unsuitable for our problem, e.g. the mutual information based selection
requires very large amounts of data. Each problem is assigned a target ordering, or class c = 1; : : : ; C, where
C = 6. Let P rc;n denote problem number n from the training dataset that is assigned class number c, c = 1; : : : ; C
and n = 1; : : : ; Nc, where Nc denotes the number of problems that are assigned class c: Thus PC
c=1 Nc = N:
1Freely available from http://cs.nyu.edu/ dejan/nonlinear/
The F-value for feature number i is computed as follows [
          <xref ref-type="bibr" rid="ref36">36</xref>
          ].
        </p>
        <p>Fi =
where fc(i) is the sample mean in class c, and f (i) the overall mean of the data:
f (i) =
c
1 XNc f (i) (P rc;n) ;
Nc n=1
f (i) = 1 XC XNc f (i)(P rc;n):</p>
        <p>N c=1 n=1
The numerator in (5) represents the between-class variability or explained variance and the denominator the
within-class variability or unexplained variance. The three features with the highest F-values were:
(5)
f65 (P r) = max0 avm;p P0 sign (d2m;p) = P1 Pp M1p Pm sign (d2m;p) ;
f46 (P r) = max0 Pp avmsign (d2m;p) (Pv0 dvm0;p)
= P</p>
        <p>p M1p Pm sign (d2m;p) (Pv0 dvm0;p) ;
f76 (P r) = av0 Pp maxm sign (d2m;p) (Pv0 dvm0;p)</p>
        <p>= Pp maxm sign (d2m;p) (Pv0 dvm0;p) :
The new features may be translated back into natural language. For example, feature 65 is the proportion
of monomials containing variable x2, averaged across all polynomials; feature 46 the sum of the degrees of
the variables in all monomials containing variable x2, averaged across all monomials and summed across all
polynomials; and feature 76 the maximum sum of the degrees of the variables in all monomials containing
variable x2, summed across all polynomials.</p>
        <p>Feature selection did not suggest to remove any features (they all contributed meaningful information), so we
proceed with our experiment using all 78.
4.6</p>
      </sec>
      <sec id="sec-4-3">
        <title>ML models</title>
        <p>
          Four of the most commonly used deterministic ML models were tuned on the training data (for details on the
methods see e.g. the textbook [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]). We also give an overview of each in [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ].
        </p>
        <p>The K</p>
        <sec id="sec-4-3-1">
          <title>Nearest Neighbours (KNN) classi er [2, x2.5].</title>
        </sec>
        <sec id="sec-4-3-2">
          <title>The Multi-Layer Perceptron (MLP) classi er [2, x2.5].</title>
        </sec>
        <sec id="sec-4-3-3">
          <title>The Decision Tree (DT) classi er [2, x14.4].</title>
          <p>The Support Vector Machine (SVM) classi er with RBF kernel [2, x6.3].</p>
          <p>Each model was trained using grid search 5-fold cross-validation, i.e. the set was randomly divided into 5 and
each possible combination of 4 parts was used to tune the model parameters, leaving the last part for tting
the hyperparameters with cross-validation, by optimising the average F-score. Grid searches were performed for
an initially large range for each hyperparameter; then gradually decreased to home into optimal values. This
lasted from a few seconds for simpler models like KNN to a few minutes for more complex models like MLP. The
optimal hyperparameters selected during cross-validation are in Table 3.
4.7</p>
        </sec>
      </sec>
      <sec id="sec-4-4">
        <title>Comparing with human made heuristics</title>
        <p>
          The ML approaches were compared in terms of prediction accuracy and resulting CAD computing time against
the two best known human constructed heuristics Brown [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] and sotd [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ] as discussed earlier. Unlike the ML,
these can end up predicting several variable orderings (i.e. when they cannot discriminate). In practice if this
were to happen the heuristic would select one randomly (or perhaps lexicographically), however that nal pick is
not meaningful. To accommodate this, for each problem, the prediction accuracy of such a heuristic is judged to
be the percentage of its predicted variable orderings that are also target orderings. The average of this percentage
over all problems in the testing dataset represents the prediction accuracy. Similarly, the computing time for
such methods was assessed as the average computing time over all predicted orderings, and it is this that is
summed up for all problems in the testing dataset.
The results are presented in Table 4. We compare the four ML models on accuracy, de ned as the percentage of
problems where they selected the optimum ordering, and also the total computation time (in seconds) for solving
all the problems with their chosen orderings. The rst two rows of the top table reproduce the results of [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ]
which used only the 11 features from Table 1, while the latter two rows are the results from the new experiment
in the present paper which has 78 features. We also compare with the two human constructed heuristics and the
outcome of a random choice between the 6 orderings (which do not change with the number of features). We
might expect a random choice to be correct one sixth of the time but it is higher as for some problems there were
multiple variable orderings with equally fast timings. The distribution of the computation times: the di erences
between the computation time of each method and the minimum computation time, given as a percentage of the
minimum time, are depicted in Figure 1.
The minimum total computing time, achieved if we select an optimal ordering for every problem, is 8 623s (the
virtual best heuristic). Choosing at random would take 30 235s, almost 4 times as much. The maximum time, if
we selected the worst ordering for every problem, is 64 534s. The K-Nearest Neighbours model (with 78 features)
achieved the shortest time of our models and heuristics, with 9 178s, only 6% more than the minimal possible.
        </p>
        <p>K-Nearest Neighbors
10
15
20
10
15
20
Multi Layer Perceptron</p>
        <p>Support Vector Machine
10
15
20
10
15
20
Brown heuristic</p>
        <p>Sotd heuristic
5 10 15</p>
        <p>Pecentage increase
from minimum time (%)
20</p>
        <p>5 10 15</p>
        <p>
          Pecentage increase
from minimum time (%)
20
Since they are not a ected by the new feature framework of the present paper the ndings on the human made
heuristics are the same as in [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ]. Of the two human-made heuristics, Brown performed the best, surprising since
the sotd heuristic has access to additional information (not just the input polynomials but also their projections).
Obtaining an ordering for a problem instance with sotd hence takes longer than for Brown or any ML model
generating an ordering with sotd for all problems in the testing dataset took over 30min. Using Brown we can
solve all problems in 10,951s, 27% more than the minimum. While sotd is only 0.7% less accurate than Brown
in identifying the best ordering, it is much slower at 11 938s or 38% more than the minimum. So, while Brown
is not much better at identifying the best, it is much better at discarding the worst!
5.3
        </p>
      </sec>
      <sec id="sec-4-5">
        <title>ML choices</title>
        <p>
          The results show that all ML approaches outperform the human constructed heuristics in terms of both
accuracy and timings. Moreover, the results show that the new algorithm for generating features leads to a clear
improvement in ML performance compared to using only a small number of human generated features in [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ].
For all four modules both accuracy has increased and computation time decreased. The best achieved time was
14% above the minimum using the original 11 features but now only 6% above with the new features.
        </p>
        <p>The computing time for all the methods lies between the best (8 623s) and the worst (64 534s). Therefore, if
we scale this time to [0; 100] so that the shortest time corresponds to 0 and the slowest to 100, then the best
human-made heuristic (Brown) lies at 4:16, and the best ML method (KNN) lies at 0:99. So using ML allows us
to be 4 times closer to the minimum possible computing time.</p>
        <p>Figure 1 shows that the human-made heuristics result in computing times that are often signi cantly larger
than 1% of the corresponding minimum time for each problem. The ML methods, on the other hand, all result
in over 1000 problems ( 75% of the testing dataset) within 1% of the minimum time.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Final Thoughts</title>
      <p>In this experiment the MLP and KNN models o ered the best performance, and a clear advance on the prior
state of the art. But we acknowledge that there is much more to do and emphasise that these are only the initial
ndings of the project and we need to see if the ndings are replicated. Planned extensions include: expanding
the dataset to problems with more variables and quanti er structure; trying di erent feature selection techniques,
and seeing if classi ers trained for the Maple CAD may be applied to other implementations.</p>
      <p>Our main result is that a great many more features can be obtained trivially from the input (i.e. without
any projection operations) than previously thought, and that these are relevant and lead to better ML choices.
Some of these are easy to express in natural language, such as the number of polynomials containing a certain
variable, but others do not have an obvious interpretation. This is important because something that is hard to
describe in natural language is unlikely to be suggested by a human as a feature, which illustrates the bene t of
our framework. This contribution to feature extraction for algebraic problems should be more widely applicable
than the CAD variable ordering decision.</p>
      <sec id="sec-5-1">
        <title>Acknowledgements</title>
        <p>This work is supported by EPSRC Project EP/R019622/1: Embedding Machine Learning within Quanti er
Elimination Procedures.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Abraham</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Abbott</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Becker</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bigatti</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brain</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Buchberger</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cimatti</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Davenport</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>England</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fontaine</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Forrest</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Griggio</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kroening</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Seiler</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sturm</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>SC2: Satis ability checking meets symbolic computation</article-title>
          .
          <source>In: Proc. CICM '16, LNCS 9791</source>
          , pp.
          <volume>28</volume>
          {
          <fpage>43</fpage>
          . Springer (
          <year>2016</year>
          ), https://doi.org/10.1007/978-3-
          <fpage>319</fpage>
          -42547-4 3
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Bishop</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <source>Pattern Recognition and Machine Learning</source>
          . Springer (
          <year>2006</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Bradford</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Davenport</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>England</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Moreno</given-names>
            <surname>Maza</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          , Wilson, D.:
          <article-title>Truth table invariant cylindrical algebraic decomposition by regular chains</article-title>
          .
          <source>In: Proc. CASC '14, LNCS 8660</source>
          , pp.
          <volume>44</volume>
          {
          <fpage>58</fpage>
          . Springer (
          <year>2014</year>
          ), http://dx.doi.
          <source>org/10.1007/978-3-319-10515-4 4</source>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Bradford</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Davenport</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>England</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Errami</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gerdt</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grigoriev</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hoyt</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kosta</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Radulescu</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sturm</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Weber</surname>
            ,
            <given-names>A.:</given-names>
          </string-name>
          <article-title>A case study on the parametric occurrence of multiple steady states</article-title>
          .
          <source>In: Proc. ISSAC '17</source>
          , pp.
          <volume>45</volume>
          {
          <fpage>52</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          (
          <year>2017</year>
          ), https://doi.org/10.1145/3087604.3087622
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Bradford</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Davenport</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>England</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>McCallum</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , Wilson, D.:
          <article-title>Truth table invariant cylindrical algebraic decomposition</article-title>
          .
          <source>J. Symb. Comp</source>
          .
          <volume>76</volume>
          ,
          <issue>1</issue>
          {
          <fpage>35</fpage>
          (
          <year>2016</year>
          ), http://dx.doi.org/10.1016/j.jsc.
          <year>2015</year>
          .
          <volume>11</volume>
          .002
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Bradford</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Davenport</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>England</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , Wilson, D.:
          <article-title>Optimising problem formulations for cylindrical algebraic decomposition</article-title>
          .
          <source>In: Intelligent Computer Mathematics, LNCS 7961</source>
          , pp.
          <volume>19</volume>
          {
          <fpage>34</fpage>
          . Springer Berlin Heidelberg (
          <year>2013</year>
          ), http://dx.doi.
          <source>org/10.1007/978-3-642-39320-4 2</source>
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Bridge</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Holden</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Paulson</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Machine learning for rst-order theorem proving</article-title>
          .
          <source>Journal of Automated Reasoning</source>
          <volume>53</volume>
          ,
          <issue>141</issue>
          {
          <fpage>172</fpage>
          (
          <year>2014</year>
          ), https://doi.org/10.1007/s10817-014-9301-5
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Brown</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          : Tutorial:
          <article-title>Cylindrical algebraic decomposition</article-title>
          , at ISSAC '
          <volume>04</volume>
          , (
          <year>2004</year>
          ), http://www.usna.edu/Users/cs/wcbrown/research/ISSAC04/handout.pdf
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Brown</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Constructing a single open cell in a cylindrical algebraic decomposition</article-title>
          .
          <source>In: Proc. ISSAC '13</source>
          , pp.
          <volume>133</volume>
          {
          <fpage>140</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          (
          <year>2013</year>
          ), https://doi.org/10.1145/2465506.2465952
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Brown</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Davenport</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>The complexity of quanti er elimination and cylindrical algebraic decomposition</article-title>
          .
          <source>In: Proc. ISSAC '07</source>
          , pp.
          <volume>54</volume>
          {
          <fpage>60</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          (
          <year>2007</year>
          ), https://doi.org/10.1145/1277548.1277557
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Carette</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          :
          <article-title>Understanding expression simpli cation</article-title>
          .
          <source>In: Proc. ISSAC '04</source>
          , pp.
          <volume>72</volume>
          {
          <fpage>79</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          (
          <year>2004</year>
          ), https://doi.org/10.1145/1005285.1005298
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Caviness</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Johnson</surname>
          </string-name>
          , J.:
          <source>Quanti er Elimination and Cylindrical Algebraic Decomposition. Texts &amp; Monographs in Symbolic Computation</source>
          , Springer-Verlag (
          <year>1998</year>
          ), https://doi.org/10.1007/978-3-
          <fpage>7091</fpage>
          -9459-1
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moreno</surname>
            <given-names>Maza</given-names>
          </string-name>
          ,
          <string-name>
            <surname>M.:</surname>
          </string-name>
          <article-title>An incremental algorithm for computing cylindrical algebraic decompositions</article-title>
          .
          <source>In: Computer Mathematics</source>
          , pp.
          <volume>199</volume>
          |
          <fpage>221</fpage>
          . Springer Berlin Heidelberg (
          <year>2014</year>
          ), https://doi.org/10.1007/978-3-
          <fpage>662</fpage>
          -43799-5 17
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moreno</surname>
            <given-names>Maza</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Xia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            ,
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <surname>L.</surname>
          </string-name>
          :
          <article-title>Computing cylindrical algebraic decomposition via triangular decomposition</article-title>
          .
          <source>In: Proc. ISSAC</source>
          '
          <volume>09</volume>
          ,
          <issue>95</issue>
          {
          <fpage>102</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          (
          <year>2009</year>
          ), https://doi.org/10.1145/1576702.1576718
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15] Collins,
          <string-name>
            <surname>G.</surname>
          </string-name>
          :
          <article-title>Quanti er elimination for real closed elds by cylindrical algebraic decomposition</article-title>
          .
          <source>In: Proc. 2nd GI Conference on Automata Theory and Formal Languages</source>
          . pp.
          <volume>134</volume>
          {
          <fpage>183</fpage>
          . Springer-Verlag (
          <article-title>reprinted in the collection</article-title>
          [
          <volume>12</volume>
          ]
          <string-name>
            <surname>)</surname>
          </string-name>
          (
          <year>1975</year>
          ), https://doi.org/10.1007/3-540-07407-4 17
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16] Collins,
          <string-name>
            <given-names>G.</given-names>
            ,
            <surname>Hong</surname>
          </string-name>
          , H.:
          <article-title>Partial cylindrical algebraic decomposition for quanti er elimination</article-title>
          .
          <source>Journal of Symbolic Computation</source>
          <volume>12</volume>
          ,
          <issue>299</issue>
          {
          <fpage>328</fpage>
          (
          <year>1991</year>
          ), https://doi.org/10.1016/S0747-
          <volume>7171</volume>
          (
          <issue>08</issue>
          )
          <fpage>80152</fpage>
          -
          <lpage>6</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Davenport</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bradford</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>England</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , Wilson,
          <string-name>
            <surname>D.</surname>
          </string-name>
          :
          <article-title>Program veri cation in the presence of complex numbers, functions with branch cuts etc</article-title>
          .
          <source>In: Proc. SYNASC '12</source>
          , pp.
          <volume>83</volume>
          {
          <fpage>88</fpage>
          .
          <string-name>
            <surname>IEEE</surname>
          </string-name>
          (
          <year>2012</year>
          ), http://dx.doi.org/10.1109/SYNASC.
          <year>2012</year>
          .68
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Dolzmann</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Seidl</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sturm</surname>
          </string-name>
          , T.:
          <article-title>E cient projection orders for CAD</article-title>
          .
          <source>In: Proc. ISSAC '04</source>
          , pp.
          <volume>111</volume>
          {
          <fpage>118</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          (
          <year>2004</year>
          ), https://doi.org/10.1145/1005285.1005303
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>England</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>Machine learning for mathematical software</article-title>
          .
          <source>In: Mathematical Software { Proc. ICMS '18. LNCS 10931</source>
          , pp.
          <volume>165</volume>
          {
          <fpage>174</fpage>
          . Springer (
          <year>2018</year>
          ), https://doi.org/10.1007/978-3-
          <fpage>319</fpage>
          -96418-8 20
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>England</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bradford</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Davenport</surname>
          </string-name>
          , J.:
          <article-title>Improving the use of equational constraints in cylindrical algebraic decomposition</article-title>
          .
          <source>In: Proc. ISSAC '15</source>
          , pp.
          <volume>165</volume>
          {
          <fpage>172</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          (
          <year>2015</year>
          ), http://dx.doi.org/10.1145/2755996.2756678
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <surname>England</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bradford</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Davenport</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          , Wilson, D.:
          <article-title>Choosing a variable ordering for truth-table invariant CAD by incremental triangular decomposition</article-title>
          .
          <source>In: Proc. ICMS '14, LNCS 8592</source>
          , pp.
          <volume>450</volume>
          {
          <fpage>457</fpage>
          . Springer (
          <year>2014</year>
          ), http://dx.doi.
          <source>org/10.1007/978-3-662-44199-2 68</source>
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <surname>England</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Davenport</surname>
          </string-name>
          , J.:
          <article-title>Experience with heuristics, benchmarks &amp; standards for cylindrical algebraic decomposition</article-title>
          .
          <source>In: Proc. SC2 '16. CEUR-WS</source>
          <year>1804</year>
          (
          <year>2016</year>
          ), http://ceur-ws.
          <source>org/</source>
          Vol-1804/
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <surname>England</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Errami</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grigoriev</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Radulescu</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sturm</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Weber</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Symbolic versus numerical computation and visualization of parameter regions for multistationarity of biological networks</article-title>
          .
          <source>In: Computer Algebra in Scienti c Computing (CASC '17)</source>
          ,
          <source>LNCS 10490</source>
          , pp.
          <volume>93</volume>
          {
          <fpage>108</fpage>
          . Springer (
          <year>2017</year>
          ), https://doi.org/10.1007/978-3-
          <fpage>319</fpage>
          -66320-3 8
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <surname>England</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Florescu</surname>
            ,
            <given-names>D.:</given-names>
          </string-name>
          <article-title>Comparing machine learning models to choose the variable ordering for cylindrical algebraic decomposition</article-title>
          . To appear
          <source>in Proc. CICM '19</source>
          (
          <string-name>
            <surname>Springer</surname>
            <given-names>LNCS</given-names>
          </string-name>
          ) (
          <year>2019</year>
          ). Preprint: https://arxiv.org/abs/
          <year>1904</year>
          .11061
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <surname>England</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , Wilson,
          <string-name>
            <given-names>D.</given-names>
            ,
            <surname>Bradford</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            ,
            <surname>Davenport</surname>
          </string-name>
          ,
          <string-name>
            <surname>J.:</surname>
          </string-name>
          <article-title>Using the Regular Chains Library to build cylindrical algebraic decompositions by projecting and lifting</article-title>
          .
          <source>In: Mathematical Software { ICMS '14. LNCS 8592</source>
          , pp.
          <volume>458</volume>
          {
          <fpage>465</fpage>
          . Springer (
          <year>2014</year>
          ), http://dx.doi.
          <source>org/10.1007/978-3-662-44199-2 69</source>
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>England</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Davenport</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Paulson</surname>
            ,
            <given-names>L.:</given-names>
          </string-name>
          <article-title>Using machine learning to decide when to precondition cylindrical algebraic decomposition with Groebner bases</article-title>
          .
          <source>In: Proc. SYNASC '16</source>
          . pp.
          <volume>45</volume>
          {
          <fpage>52</fpage>
          .
          <string-name>
            <surname>IEEE</surname>
          </string-name>
          (
          <year>2016</year>
          ), https://doi.org/10.1109/SYNASC.
          <year>2016</year>
          .020
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>England</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , Wilson,
          <string-name>
            <given-names>D.</given-names>
            ,
            <surname>Bridge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>Davenport</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>Paulson</surname>
          </string-name>
          ,
          <string-name>
            <surname>L.:</surname>
          </string-name>
          <article-title>Using machine learning to improve cylindrical algebraic decomposition</article-title>
          .
          <source>Mathematics in Computer Science</source>
          , Volume to be assigned, 28 pages (
          <year>2019</year>
          ), https://doi.org/10.1007/s11786-019-00394-8
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>England</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , Wilson,
          <string-name>
            <given-names>D.</given-names>
            ,
            <surname>Davenport</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>Paulson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            ,
            <surname>Bridge</surname>
          </string-name>
          ,
          <string-name>
            <surname>J.:</surname>
          </string-name>
          <article-title>Applying machine learning to the problem of choosing a heuristic to select the variable ordering for cylindrical algebraic decomposition</article-title>
          .
          <source>In: Intelligent Computer Mathematics, LNAI 8543</source>
          , pp.
          <volume>92</volume>
          {
          <fpage>107</fpage>
          . Springer International (
          <year>2014</year>
          ), http://dx.doi.
          <source>org/10.1007/978-3-319-08434-3 8</source>
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <surname>Iwane</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yanami</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Anai</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yokoyama</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>An e ective implementation of a symbolic-numeric cylindrical algebraic decomposition for quanti er elimination</article-title>
          .
          <source>In: Proc. SNC '09</source>
          , pp.
          <volume>55</volume>
          {
          <fpage>64</fpage>
          . SNC '
          <volume>09</volume>
          (
          <year>2009</year>
          ), https://doi.org/10.1145/1577190.1577203
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <surname>Jovanovic</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>de Moura</surname>
          </string-name>
          , L.:
          <article-title>Solving non-linear arithmetic</article-title>
          . In: Gramlich,
          <string-name>
            <given-names>B.</given-names>
            ,
            <surname>Miller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            ,
            <surname>Sattler</surname>
          </string-name>
          ,
          <string-name>
            <surname>U</surname>
          </string-name>
          . (eds.)
          <source>Automated Reasoning Proc. IJCAR '12, LNCS 7364</source>
          , pp.
          <volume>339</volume>
          {
          <fpage>354</fpage>
          . Springer (
          <year>2012</year>
          ), https://doi.org/10.1007/978-3-
          <fpage>642</fpage>
          -31365-3 27
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <surname>Kobayashi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Iwane</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Matsuzaki</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Anai</surname>
          </string-name>
          , H.:
          <article-title>E cient subformula orders for real quanti er elimination of non-prenex formulas</article-title>
          .
          <source>In: Proc. MACIS '15, LNCS 9582</source>
          , pp.
          <volume>236</volume>
          {
          <fpage>251</fpage>
          . Springer International Publishing (
          <year>2016</year>
          ), https://doi.org/10.1007/978-3-
          <fpage>319</fpage>
          -32859-1 21
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32] Kuhlwein,
          <string-name>
            <given-names>D.</given-names>
            ,
            <surname>Blanchette</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>Kaliszyk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            ,
            <surname>Urban</surname>
          </string-name>
          ,
          <string-name>
            <surname>J.:</surname>
          </string-name>
          <article-title>MaSh: Machine learning for sledgehammer</article-title>
          .
          <source>In: Interactive Theorem Proving, LNCS 7998</source>
          , pp.
          <volume>35</volume>
          {
          <fpage>50</fpage>
          . Springer Berlin Heidelberg (
          <year>2013</year>
          ), https://doi.org/10.1007/978-3-
          <fpage>642</fpage>
          -39634-2 6
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [33]
          <string-name>
            <surname>Kuipers</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ueda</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Vermaseren</surname>
            ,
            <given-names>J.A.M.:</given-names>
          </string-name>
          <article-title>Code optimization in FORM</article-title>
          .
          <source>Computer Physics Communications</source>
          ,
          <volume>189</volume>
          , 1{
          <fpage>19</fpage>
          (
          <year>2015</year>
          ), https://doi.org/10.1016/j.cpc.
          <year>2014</year>
          .
          <volume>08</volume>
          .008
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          [34]
          <string-name>
            <surname>Liang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Hari</given-names>
            <surname>Govind</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            ,
            <surname>Poupart</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Czarnecki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            ,
            <surname>Ganesh</surname>
          </string-name>
          ,
          <string-name>
            <surname>V.</surname>
          </string-name>
          :
          <article-title>An empirical study of branching heuristics through the lens of global learning rate</article-title>
          .
          <source>In: Proc. SAT '17, LNCS 10491</source>
          , pp.
          <volume>119</volume>
          {
          <fpage>135</fpage>
          . Springer (
          <year>2017</year>
          ), https://doi.org/10.1007/978-3-
          <fpage>319</fpage>
          -66263-3 8
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          [35]
          <string-name>
            <surname>McCallum</surname>
            ,
            <given-names>S.:</given-names>
          </string-name>
          <article-title>An improved projection operation for cylindrical algebraic decomposition</article-title>
          .
          <source>In: [12]</source>
          , pp.
          <volume>242</volume>
          {
          <fpage>268</fpage>
          . (
          <year>1998</year>
          ), https://doi.org/10.1007/978-3-
          <fpage>7091</fpage>
          -9459-1 12
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          [36]
          <string-name>
            <surname>Markowski</surname>
            ,
            <given-names>C.A.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Markowski</surname>
            ,
            <given-names>E.P.</given-names>
          </string-name>
          :
          <article-title>Conditions for the e ectiveness of a preliminary test of variance</article-title>
          .
          <source>The American Statistician</source>
          ,
          <volume>44</volume>
          :
          <fpage>4</fpage>
          ,
          <issue>322</issue>
          {
          <fpage>326</fpage>
          (
          <year>1990</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          [37]
          <string-name>
            <surname>McCallum</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Parusiniski</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Paunescu</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Validity proof of Lazard's method for CAD construction</article-title>
          .
          <source>Journal of Symbolic Computation</source>
          <volume>92</volume>
          ,
          <issue>52</issue>
          {
          <fpage>69</fpage>
          (
          <year>2019</year>
          ), https://doi.org/10.1016/j.jsc.
          <year>2017</year>
          .
          <volume>12</volume>
          .002
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          [38]
          <string-name>
            <surname>Mulligan</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bradford</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Davenport</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>England</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tonks</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          :
          <article-title>Non-linear real arithmetic benchmarks derived from automated reasoning in economics</article-title>
          .
          <source>In: Proc. SC2 '18</source>
          , pp.
          <volume>48</volume>
          {
          <fpage>60</fpage>
          . CEUR-WS
          <volume>2189</volume>
          (
          <year>2018</year>
          ), http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>2189</volume>
          /
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          [39]
          <string-name>
            <surname>Mulligan</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Davenport</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>England</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>TheoryGuru: A Mathematica package to apply quanti er elimination technology to economics</article-title>
          .
          <source>In: Mathematical Software { Proc. ICMS '18. LNCS 10931</source>
          , pp.
          <volume>369</volume>
          {
          <fpage>378</fpage>
          . Springer (
          <year>2018</year>
          ), https://doi.org/10.1007/978-3-
          <fpage>319</fpage>
          -96418-8 44
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          [40]
          <string-name>
            <surname>Pedregosa</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Varoquaux</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gramfort</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Michel</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Thirion</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grisel</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Blondel</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Prettenhofer</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Weiss</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dubourg</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vanderplas</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Passos</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cournapeau</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brucher</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Perrot</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Duchesnay</surname>
          </string-name>
          , E.:
          <article-title>Scikit-learn: Machine learning in Python</article-title>
          .
          <source>Journal of Machine Learning Research</source>
          <volume>12</volume>
          ,
          <volume>2825</volume>
          {
          <fpage>2830</fpage>
          (
          <year>2011</year>
          ), http://www.jmlr.org/papers/v12/pedregosa11a.html
        </mixed-citation>
      </ref>
      <ref id="ref41">
        <mixed-citation>
          [41]
          <string-name>
            <surname>Strzebonski</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Cylindrical algebraic decomposition using validated numerics</article-title>
          .
          <source>Journal of Symbolic Computation</source>
          <volume>41</volume>
          (
          <issue>9</issue>
          ),
          <volume>1021</volume>
          {
          <fpage>1038</fpage>
          (
          <year>2006</year>
          ), https://doi.org/10.1016/j.jsc.
          <year>2006</year>
          .
          <volume>06</volume>
          .004
        </mixed-citation>
      </ref>
      <ref id="ref42">
        <mixed-citation>
          [42]
          <string-name>
            <surname>Sturm</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>New domains for applied quanti er elimination</article-title>
          . In: Computer Algebra in Scienti c Computing,
          <source>LNCS 4194</source>
          , pp.
          <volume>295</volume>
          {
          <fpage>301</fpage>
          . Springer (
          <year>2006</year>
          ), https://doi.org/10.1007/11870814 25
        </mixed-citation>
      </ref>
      <ref id="ref43">
        <mixed-citation>
          [43]
          <string-name>
            <surname>Urban</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>MaLARea: A metasystem for automated reasoning in large theories</article-title>
          .
          <source>In: Proc. ESARLT '07, CEUR-WS 257</source>
          , p.
          <fpage>14</fpage>
          .
          <string-name>
            <surname>CEUR-WS</surname>
          </string-name>
          (
          <year>2007</year>
          ), http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>257</volume>
          /
        </mixed-citation>
      </ref>
      <ref id="ref44">
        <mixed-citation>
          [44] Wilson,
          <string-name>
            <given-names>D.</given-names>
            ,
            <surname>Bradford</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            ,
            <surname>Davenport</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>England</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          :
          <article-title>Cylindrical algebraic sub-decompositions</article-title>
          .
          <source>Mathematics in Computer Science</source>
          <volume>8</volume>
          ,
          <issue>263</issue>
          {
          <fpage>288</fpage>
          (
          <year>2014</year>
          ), http://dx.doi.org/10.1007/s11786-014-0191-z
        </mixed-citation>
      </ref>
      <ref id="ref45">
        <mixed-citation>
          [45] Wilson,
          <string-name>
            <given-names>D.</given-names>
            ,
            <surname>Davenport</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>England</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Bradford</surname>
          </string-name>
          ,
          <string-name>
            <surname>R.:</surname>
          </string-name>
          <article-title>A \piano movers" problem reformulated</article-title>
          .
          <source>In: Proc. SYNASC '13</source>
          , pp.
          <volume>53</volume>
          {
          <fpage>60</fpage>
          .
          <string-name>
            <surname>IEEE</surname>
          </string-name>
          (
          <year>2013</year>
          ), http://dx.doi.org/10.1109/SYNASC.
          <year>2013</year>
          .14
        </mixed-citation>
      </ref>
      <ref id="ref46">
        <mixed-citation>
          [46]
          <string-name>
            <surname>Xu</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hutter</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hoos</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Leyton-Brown</surname>
          </string-name>
          , K.:
          <article-title>SATzilla: Portfolio-based algorithm selection for SAT</article-title>
          .
          <source>J. Arti cial Intelligence Research</source>
          <volume>32</volume>
          ,
          <volume>565</volume>
          {
          <fpage>606</fpage>
          (
          <year>2008</year>
          ), https://doi.org/10.1613/jair.2490
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>