<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Cautious Specialization of Strategy Schedules (Extended Abstract)</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Filip</forename><surname>Bártek</surname></persName>
							<email>filip.bartek@cvut.cz</email>
							<affiliation key="aff0">
								<orgName type="department">Robotics and Cybernetics</orgName>
								<orgName type="institution">Czech Technical University in Prague -Czech Institute of Informatics</orgName>
								<address>
									<addrLine>Jugoslávských partyzánů 1580/3</addrLine>
									<postCode>160 00</postCode>
									<settlement>Prague 6 -Dejvice</settlement>
									<country key="CZ">Czech Republic</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="department">Prague -Faculty of Electrical Engineering</orgName>
								<orgName type="institution">Czech Technical University</orgName>
								<address>
									<addrLine>Technická 2, 166 27</addrLine>
									<settlement>Prague 6 -Dejvice</settlement>
									<country key="CZ">Czech Republic</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Karel</forename><surname>Chvalovský</surname></persName>
							<email>karel.chvalovsky@cvut.cz</email>
							<affiliation key="aff0">
								<orgName type="department">Robotics and Cybernetics</orgName>
								<orgName type="institution">Czech Technical University in Prague -Czech Institute of Informatics</orgName>
								<address>
									<addrLine>Jugoslávských partyzánů 1580/3</addrLine>
									<postCode>160 00</postCode>
									<settlement>Prague 6 -Dejvice</settlement>
									<country key="CZ">Czech Republic</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Martin</forename><surname>Suda</surname></persName>
							<email>martin.suda@cvut.cz</email>
							<affiliation key="aff0">
								<orgName type="department">Robotics and Cybernetics</orgName>
								<orgName type="institution">Czech Technical University in Prague -Czech Institute of Informatics</orgName>
								<address>
									<addrLine>Jugoslávských partyzánů 1580/3</addrLine>
									<postCode>160 00</postCode>
									<settlement>Prague 6 -Dejvice</settlement>
									<country key="CZ">Czech Republic</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Cautious Specialization of Strategy Schedules (Extended Abstract)</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">DD7857FA567FE4B4A11E5D30A13CAF07</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T19:18+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>strategy schedules, schedule specialization, regularization, Vampire (M. Suda) 0000-0002-1822-2651 (F. Bártek)</term>
					<term>0000-0002-0541-3889 (K. Chvalovský)</term>
					<term>0000-0003-0989-5800 (M. Suda)</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Combining theorem proving strategies into schedules is a well-known recipe for improving prover performance, as a well-chosen set of complementary strategies in practice covers many more problems within a predetermined time budget than a single strategy could. A strategy schedule can be a monolithic sequence, but may also consist of multiple schedule branches, intended to be selected based on certain problem's features and thus to specialize in solving the corresponding problem classes. When allowing schedule branching, there is an intuitive trade-off between covering the known (training) problems quickly, by splitting them into many classes, and the generalization of the obtained branching schedule, i.e., its ability to solve previously unseen (test) problems.</p><p>In this paper, we attempt to shed more light on this trade-off. Starting off with a large set of proving strategies of the automatic theorem prover Vampire evaluated on the first-order part of the TPTP library, we report on two preliminary experiments with building branching schedules while keeping track of how well they generalize to test problems. We hope to attract more attention to the exciting topic of schedule construction by this work-in-progress report.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Proof search in most automatic theorem provers (ATPs) relies on parameterized heuristics. Strong configurations (strategies) of an ATP are, typically, specialized to certain classes of input problems -no configuration performs well on all possible problems of interest. Selecting a strong strategy for a given input problem is, in general, hard. This is namely due to the chaotic behavior of the provers <ref type="bibr" target="#b0">[1]</ref>: Using a fixed strategy, a small change in the input problem may lead to a substantially different proof search.</p><p>Typically, if a strategy solves a problem, the solution is reached in a relatively short time <ref type="bibr" target="#b1">[2]</ref>. In contexts where sufficient time or parallelism is available, running a portfolio of diverse strategies is a pragmatic approach to increase the power of a prover. Such portfolio, or schedule in case a time limit is assigned to each of the strategies, may be either hand-crafted <ref type="bibr" target="#b2">[3]</ref> or constructed automatically <ref type="bibr" target="#b3">[4,</ref><ref type="bibr" target="#b4">5]</ref>.</p><p>Further gain in performance can be achieved by specializing the schedule: We partition the possible input problems into classes and construct a specialized schedule for each class. We prefer the classes to be homogeneous in terms of the problems' response to strategies, so that a schedule specialized for a class needs less time to cover the same percentage of problems. Increasing the granularity of the classes increases the specialization and the associated potential performance gain.</p><p>Decision trees <ref type="bibr" target="#b5">[6]</ref> provide a suitable architecture to capture such gradual specialization. A set of features, such as the number of atoms, the presence of certain syntactic traits, or the membership in standard logical fragments, are computed for the input problem. A binary tree is then traversed from the root to a leaf. Each internal node is labeled with a criterion that partitions the domain of one of the features into two parts. When a node is traversed, the next node is selected depending on the corresponding feature of the problem. The schedule inhabiting the final leaf is then used to attempt to solve the problem.</p><p>The ATP Vampire <ref type="bibr" target="#b6">[7]</ref> won the prestigious First-Order Form theorems (FOF) division of the CADE ATP System Competition (CASC) <ref type="bibr" target="#b7">[8]</ref> every year from 2002 to 2023 <ref type="bibr" target="#b8">[9]</ref>. Arguably, this continued success is partially thanks to the famous "CASC mode" of the prover and the corresponding use of strong strategy schedules. The schedules were constructed by Andrei Voronkov with the help of a system called Spider <ref type="bibr" target="#b9">[10]</ref> and arranged in a shallow decision tree for specialization. <ref type="foot" target="#foot_0">1</ref>When the specialization problem classes are chosen in a way that optimizes the performance on a training set of problems, we run the risk of overfitting, that is increasing performance on the training problems at the cost of decreasing performance on unseen problems. Regularization is the process of reducing the expressive power of a machine-learned model to prevent overfitting.</p><p>In our previous work <ref type="bibr" target="#b3">[4]</ref>, we studied the regularization of monolithic schedule construction, i.e., the case where the algorithm's output is a single sequence with no specialization to problem features. We discovered and evaluated more than a thousand Vampire strategies targeting the first-order fragment of the TPTP library <ref type="bibr" target="#b10">[11]</ref>, which are now available as an independent data set for experimentation <ref type="bibr" target="#b11">[12]</ref>.</p><p>In this work, we use the same data set to start shedding some light on regularization during schedule specialization. Namely, after brief preliminaries (Sect. 2), we first discuss (and measure) how a single branching criterion (i.e., a single problem feature) can affect the generalization of the simplest branching schedule with just one level of branching (Sect. 3). We complement this view by an experiment with gradient boosted trees (Sect. 4), which explores the opposite extreme case where every problem canbased on its features -be, in principle, assigned its unique schedule. Our paper also briefly overviews related work (Sect. 5) and offers several concluding remarks (Sect. 6).</p><p>We are far from any definite answer on how specialization in schedule construction should be best done. Yet we hope to spark more interest in this exciting topic by this work-in-progress report.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Preliminaries</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Training and evaluation data</head><p>In our previous work <ref type="bibr" target="#b3">[4]</ref>, we generated a collection of 1096 strong and mutually complementary strategies for the ATP Vampire. We evaluated the performance of each of these strategies on a set of 7866 first-order logic (FOL) problems from TPTP <ref type="bibr" target="#b10">[11]</ref> v8.2.0. The results of these evaluations constitute a dataset suitable for experimenting with schedule construction and specialization <ref type="bibr" target="#b11">[12]</ref>.</p><p>We denote the set of strategies as 𝑆 (|𝑆| = 1096), the set of problems as 𝑃 (|𝑃 | = 7866), and the evaluation matrix as 𝐸 𝑠 𝑝 : 𝑆 × 𝑃 → N ∪ {∞}. The value 𝐸 𝑠 𝑝 is the time at which strategy 𝑠 solves problem 𝑝, or ∞ in case strategy 𝑠 failed to solve problem 𝑝 within the time limit.</p><p>The time is measured in whole CPU megainstructions (Mi), where 1 Mi = 2 20 instructions reported by the tool perf and 2000 Mi take approximately 1 second of wallclock time on our hardware.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Vampire's CASC-mode problem features</head><p>For our schedule specialization experiments, we use problem features already available in Vampire and used there by the CASC mode. This feature set consists mainly of various syntactic counters (the number of: goal clauses, axiom clauses, Horn clauses, equational clauses, etc.), most prominent of which is the number of atoms (roughly corresponding to problem size); problem class according to TPTP classification (NEQ, HEQ, . . . , EPR, UEQ); and a number of bitfields checking for properties such as "has 𝑋 = 𝑌 ", "has functional definitions", or "has inequality resolvable with deletion". These features are not included in the mentioned dataset <ref type="bibr" target="#b11">[12]</ref>, but can be obtained easily by running a modified version of Vampire<ref type="foot" target="#foot_1">2</ref> in "profile mode" via --mode profile &lt;problem_name&gt;.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.">Strategy schedules</head><p>A strategy schedule is a mapping 𝑡 𝑠 : 𝑆 → N. The value 𝑡 𝑠 is the time limit assigned to strategy 𝑠. A schedule may be used to attack an arbitrary problem with Vampire: The strategies are run sequentially with the assigned time limits, each attempting to solve the problem, until a solution is found or every strategy has depleted its allocated time. In the rest of this work, instead of running the prover, we use the evaluation matrix to simulate the evaluation of schedules, restricting our scope to problems in 𝑃 .</p><p>Similarly to strategies, schedules are typically evaluated in a time-constrained setting. Schedule 𝑡 𝑠 satisfies (time) budget 𝑇 ∈ N if and only if ∑︀ 𝑠∈𝑆 𝑡 𝑠 ≤ 𝑇 . In our previous work <ref type="bibr" target="#b3">[4]</ref>, we proposed a greedy algorithm that constructs a strategy schedule from an evaluation matrix and a budget. The algorithm allows finding a relatively strong (with respect to the evaluation matrix) schedule satisfying the budget quickly.</p><p>A strategy schedule recommender produces a schedule for an arbitrary input problem. To evaluate a schedule recommender on a set of problems, we estimate, using the evaluation matrix, the proportion of problems solved by the respective schedules produced by the recommender.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.4.">Overfitting and regularization</head><p>Extreme schedule specialization is discouraged in CASC <ref type="bibr" target="#b12">[13]</ref>:</p><p>All techniques used must be general purpose, and expected to extend usefully to new unseen problems. [...] If machine learning procedures are used to tune a system, the learning must ensure that sufficient generalization is obtained so that no there is no specialization to individual problems.</p><p>In machine learning (ML), it is common for a trained system to perform better on the data it was trained on than on previously unseen data. The system is said to overfit to the training data. In most applications, we are ultimately interested in optimizing the performance on unseen data, so overfitting is undesirable. A system that does not overfit is said to generalize (to unseen data).</p><p>To evaluate generalization of our ML training procedures, we repeatedly randomly split the set of problems 𝑃 into a training set (80 %) and a test set (20 %), train the system on the training set, and evaluate the trained system on the test set. The final score is the mean of the per-split test scores (in our case, schedule recommender success rates). This way, we estimate the performance on unseen data.</p><p>Moreover, before training we exclude strategies that have their "witness problem" in the test set. Roughly speaking, this prunes the set of strategies only to those that could have been discovered if a given test set was fixed beforehand for the strategy discovery phase. See [4, Section 6] for more details.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Single branching point</head><p>Creating a schedule can be seen as an act of "covering" the training problems with the help of the known solutions of the previously evaluated strategies. It is clear that if many strategies need to be involved with long running times in order to eventually cover every training problem, the running time of the whole schedule will need to be high. <ref type="foot" target="#foot_2">3</ref> Schedule specialization comes to rescue to reduce this running time, as it allows the covering process to focus, within each branch, on a different subset of the training problems. In a nutshell, on each branch, there is less covering work to do and during evaluation, only the "right" branch is always used.</p><p>However, from the perspective of generalization, i.e., the effort of ensuring that our specializing schedule performs well also on problems unseen in training, splitting the work into branches may be detrimental. Indeed, if many training problems end up in their respective subset for mostly a random  reason, the chances of a similar testing problem (that could be solved by the same strategy) being classified into the same subset and thus subjected to the same schedule branch are low.</p><p>In this experiment, we set out to demonstrate this phenomenon at its extreme. Our method of covering is the greedy algorithm presented in previous work <ref type="bibr" target="#b3">[4]</ref> and we run it repetitively over random 80 : 20 train/test splits of our problem set 𝑃 . We compare the average performance of a monolithic (non-branching) schedule, a schedule branching into three groups based on the number of problem's atoms, and a schedule branching into three random equally-sized groups. For the branching on atoms, we split the problems into small (number of atoms &lt; 750), large (number of atoms &gt; 2900) and medium (the rest). The cutoff points were chosen not to split the problems into equally-sized subsets, but by aiming for a split for which each subset can be fully covered within approximately the same budget.</p><p>The results are presented in cactus plot form in Fig. <ref type="figure" target="#fig_1">1</ref>. The results of the monolithic schedule building (1-Train/1-Test) are the same in both the left and the right plot (and roughly correspond to Fig. <ref type="figure">7</ref> in Appendix E of our previous work <ref type="bibr" target="#b3">[4]</ref>). The monolithic schedule requires approximately 2 million Mi to cover all the training problems and the test performance converges to around 93 % of the training one. The train performance of both the random 3-split (3-Train) and the atom split (Atom-Train) are at all times higher than that of the monolithic schedule until they (necessarily) converge to the same value and stabilize. (The atom split tends to finish its covering job a bit earlier than the random split.) However, the pictures for the test performance of the branching schedules differ. While splitting on atoms improves over monolithic even in test performance (up until around the 1 million Mi mark), and for the lower budgets even improves over the monolithic's train performance, the test performance of the random 3-split (3-Test) is at all times worse.</p><p>It is perhaps not surprising that random features are useless from the perspective of branching schedule generalization, however, we find the "backwards version" of this perspective to be quite revealing, namely the realization that even reasonable-sounding problem features may differ in how useful (useless, or even detrimental) they could be in this regard. Schedule construction is sometimes organized by first fixing the problem groups based on features and only then looking for strategies in each group separately (one can save computational resources by not having to evaluate each strategy on every training problem; evaluating just within the relevant group would be enough). However, if such initial split is only based on educated guessing of what problems reasonably belong together, opportunities for better (or even the best possible) generalization can be missed.</p><p>Ideally, we would like to come up with generalization-conducive splits of the problems beforehand, by just studying our evaluation matrix and asking a question such as "What could be a nice, easy-todelineate subset of problems for which certain strategies work well together and complement each other, while elsewhere they are mostly useless?" We remark that such splits might be defined by other properties than those expressible by our current fixed feature set. Discovering such new properties that would still be easy to compute (and thus easy to understand) would be an interesting instance of the explainable AI endeavour. At this moment, however, we do not have a good way of doing this and leave this research direction for future work.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Gradient boosting</head><p>Unlike in the previous section, where a hand-crafted split based on the number of atoms occurring in the problem is discussed, we want to produce finer splits of the problems based on Vampire's CASC-mode problem features here. In other words, we want to learn more complicated decision trees, or as in our case ensembles of small decision trees. Hence we want to train a model that for a given problem 𝑝 produces a schedule 𝑡 1 , . . . , 𝑡 |𝑆| , where ∑︀ 𝑠∈𝑆 𝑡 𝑠 ≤ 𝑇 , such that there is a strategy 𝑠 ∈ 𝑆 that solves 𝑝 ∈ 𝑃 in time at most 𝑡 𝑠 .</p><p>Although we want to obtain schedules, we may characterize them by probability distributions. A discrete probability mass function p : 𝑆 → [0, 1] defines a schedule by setting 𝑡 𝑠 = 𝑇 • p 𝑠 . Hence we can use standard approaches for multi-class classification, where usually we get a probability distribution over possible classes and select the class as the one with the highest estimated probability. Here, however, we take the estimated probability distribution and compute a schedule based on it.</p><p>This of course immediately leads to the question of what is the true probability distribution we want to estimate. One possible approach is to assign, for example, probability 1 to the best strategy for the problem. However, we likely need a more fine-grained approach as we are interested in generalization. First, we are happy to split the budget among more strategies, as long as at least one of them solves the problem. Second, we do not require that the problem be solved by the best available strategy.</p><p>Hence instead of having, for each (training) problem, a fixed true distribution that we want to learn, we may produce a "true" distribution iteratively during the training based on our previous estimates and change it as necessary. For example, we have a training problem 𝑝 and an estimated probability p 1 , . . . , p |𝑆| for it. We may want to produce a "true" distribution for 𝑝 by assigning p 𝑖 = 0 if the strategy 𝑖 does not solve the problem in the given budget 𝑇 , and hence increase p 𝑗 if the strategy 𝑗 solves the problem. In other words, we want to split the budget only among strategies that can possibly solve the problem. Moreover, we can boost the strategy that is closest to solving the problem; the strategy requiring the smallest increase in the estimated budget to solve the problem. And there are many other ways how to get a new "true" distribution. It is worth noticing that we may obtain this "true" distribution in different ways: one way for those training examples that our model predicts correctly (estimated schedule solves the problem) and another for those that are predicted incorrectly.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Model</head><p>Gradient boosting for decision trees, see, e.g., <ref type="bibr" target="#b13">[14]</ref>, fits nicely into this picture. Loosely speaking, it combines weak models that iteratively improve their predictions by correcting the previous mistakes. In particular, we learn decision trees that split our problems into different subsets based on the static Vampire's features. Or, more precisely, in each iteration, we learn |𝑆| regression trees (multi-class problem); one per strategy. When we want to estimate the probability distribution after 𝑖 iterations, we combine the output of all 𝑖 regression trees learned for each strategy, and get the probability distribution by the softmax function. In the next iteration, we fix the "true" distribution based on the newly estimated distribution and iterate this process until we get a satisfactory model or reach the limit on the number of iterations. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Initial experiments</head><p>We use LightGBM <ref type="bibr" target="#b14">[15]</ref>, a framework for gradient boosting decision trees, with custom objective functions. We get probability estimates from our models by the softmax function and use the crossentropy between predicted probabilities and "true" probabilities as the loss function.</p><p>We report some preliminary results, as we have not performed much of hyper-parameter tuning. <ref type="foot" target="#foot_3">4</ref>We use the learning rate 0.05, the maximal number of bins that feature values are bucketed in is 8, the number of boosting iterations is 500, and, most importantly, we restrict the number of leaves to 2; hence we use decision stumps as weak learners.</p><p>Our custom objective functions create "true" probabilities p ′ 1 , . . . , p ′ |𝑆| from the estimated probabilities p 1 , . . . , p |𝑆| for a problem 𝑝. We have tried various combinations of the following procedures: We train our models using 10 random 80:20 train/test splits over the dataset described in Section 2.1. In Table <ref type="table" target="#tab_0">1</ref>, we report the success rate. The best iteration is selected based on the number of solved problems on the training set. For shorter budgets, the gradient boosting model significantly outperforms the simple split into small-medium-large subsets based on the number of atoms in the problem. As expected, the gap is narrowing as the budget is increasing.</p><p>Interestingly, when we inspect the produced schedules, the majority of problems (both on the train and test sets) are solved by multiple strategies in the schedule. Moreover, it is unlikely that a problem is solved by the overall best available strategy for the problem, supporting our idea behind the careful creation of "true" probabilities.</p><p>As the "true" probability that we aim to estimate evolves, it is possible that the majority of trees learned do not offset incorrect estimates from previous iterations but rather compensate for the changing "true" probability. Note that we often add the same tree (with different leaf values) multiple times.</p><p>These initial experiments show that it is possible to automatically train a model that both leverages the available feature space and keeps a reasonable level of generalization.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Related work</head><p>Strategy schedule specialization is closely related to strategy selection (also referred to as "heuristic selection"), a special case of algorithm selection <ref type="bibr" target="#b15">[16,</ref><ref type="bibr" target="#b16">17]</ref>. Automatic mode in the E prover <ref type="bibr" target="#b17">[18]</ref> uses hand-crafted problem classes to select a strategy for the input problem <ref type="bibr" target="#b18">[19]</ref>. Strategy selection for E has also been approached by machine learning <ref type="bibr" target="#b19">[20]</ref>. Given an input problem, MaLeS <ref type="bibr" target="#b18">[19]</ref> predicts solving time for all strategies in a portfolio and selects the strategy that minimizes the predicted time. Similarly, Grackle <ref type="bibr" target="#b20">[21]</ref> selects a strategy using a trained LightGBM model that predicts a ranking of strategies in a pre-computed portfolio.</p><p>Strategy schedules have been constructed manually in Gandalf <ref type="bibr" target="#b2">[3]</ref> and automatically in E-SETHEO <ref type="bibr" target="#b21">[22]</ref>, Spider for Vampire <ref type="bibr" target="#b9">[10]</ref>, CPhydra <ref type="bibr" target="#b22">[23]</ref>, BliStrTune <ref type="bibr" target="#b23">[24]</ref>, HOS-ML <ref type="bibr" target="#b4">[5]</ref>, and in our previous work <ref type="bibr" target="#b3">[4]</ref>. Given an input problem, HOS-ML <ref type="bibr" target="#b4">[5]</ref> selects a strategy schedule automatically. The problem's static features are used to identify a problem class, for which a schedule was optimized during the training. CPhydra <ref type="bibr" target="#b22">[23]</ref> identifies 10 training problems that are the most similar to the input problem and optimizes a schedule for these problems using a pre-computed evaluation matrix.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusion</head><p>Working with a dataset of 1096 Vampire strategies, which were evaluated on the first-order part of TPTP, and a selection of syntactic problem features, which Vampire computes efficiently for its input problem at startup, we presented two experiments with strategy schedule construction and schedule specialization. In the first, we noticed that the selection of the splitting features not only influences how fast a specialized schedule can cover the training problems, but, more interestingly, how well it can generalize to previously unseen problems. In the second, we demonstrated that a good performance can be achieved by using gradient boosted decision trees. The second approach is interesting in that it departs from the concept of problem covering and instead develops a schedule in the form of a probability distribution over the available strategies. This requires the use of a custom target distribution which develops during the learning process.</p><p>There are many avenues for future research. Our next target for experimentation is the grow-andprune technique for decision tree construction and regularization <ref type="bibr" target="#b5">[6]</ref> and its adaptation to cautious schedule specialization. However, we would also be more than happy for others to join us in experimenting with our strategy data <ref type="bibr" target="#b11">[12]</ref> to see which method could eventually lead to the most useful approach.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Monolithic ("1") vs single branching schedule performance. Problems split into three random subsets on the left ("3") and into small-medium-large subsets based on the number of atoms ("Atoms") on the right. Showing averages over 100 random 80 : 20 train/test split runs. Shading marks the zone of one standard deviation around the averages (and got four times larger for the Test runs compared to the Train runs just from the effect of normalizing the averages from the unequal train/test set sizes to the reported fractions).</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head></head><label></label><figDesc>(a) p ′ 𝑖 = 0 if the strategy 𝑖 does not solve 𝑝 in the time budget 𝑇 , otherwise p ′ 𝑗 = p 𝑗 , (b) p ′ 𝑖 = 1 if the strategy 𝑖 is the best available strategy 5 and p ′ 𝑗 = 0 for 𝑖 ̸ = 𝑗, (c) boost 6 the best available strategy, and then L1-normalize p ′ 1 , . . . , p ′ |𝑆| . It is even possible to combine different ways how to create "true" probabilities based on whether the predicted schedule solves the training problem or not. Surprisingly, (a) works quite well even though it can get stuck in a local minima. Hence, for simplicity, all the subsequent results use this objective function. It takes roughly 10 s, using one core, to train one iteration of the model with our non-optimized custom objective function written in Python.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>We compare the success rate of the split into small-medium-large subsets based on the number of atoms ("Atoms", see Sect. 3) and the schedule produced by gradient boosting.</figDesc><table><row><cell>Budget</cell><cell>Atoms Train Test</cell><cell cols="2">Gradient boosting Train Test</cell></row><row><cell>1000</cell><cell cols="2">0.5297 0.5113 0.6196</cell><cell>0.5660</cell></row><row><cell>10000</cell><cell cols="2">0.6965 0.6547 0.7650</cell><cell>0.6998</cell></row><row><cell cols="3">100000 0.8219 0.7519 0.8358</cell><cell>0.7807</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">E.g., https://github.com/vprover/vampire/blob/c7564c1d65020771079f29787ca2d5d7743f5d6a/CASC/Schedules.cpp#L5676.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">https://github.com/vprover/vampire/tree/ourProfile</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">Here we conceptually depart from the position where there is a fixed budget and we are trying to solve as many problems as possible within that budget. This is because we aim, in this section, to present an analysis which answers a question for all budgets at once. We can do it thanks to the anytime nature of the mentioned greedy algorithm we employ here<ref type="bibr" target="#b3">[4]</ref>.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_3">The performance of LightGBM is sensitive to choices of hyper-parameters. Moreover, our custom objective functions make these choices even more significant.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_4">There are many possible definitions here. For example, we have experimented with selecting the strategy that has the maximal difference between the time assigned to the strategy and the time necessary to solve the problem by the strategy.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_5">We tried two variants. First, in combination with (a), we split "saved time" not among all strategies that solve the problem, but we only boost the best available strategy. Second, we increase the allocation to the best available strategy just enough for it to solve the problem.</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>The work on this paper was supported by the Czech Science Foundation grant 24-12759S and the project RICAIP no. 857306 under the EU-H2020 programme.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Vampire getting noisy: Will random bits help conquer chaos? (system description)</title>
		<author>
			<persName><forename type="first">M</forename><surname>Suda</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-031-10769-6_38</idno>
		<idno>doi:</idno>
		<ptr target="10.1007/978-3-031-10769-6\_38" />
	</analytic>
	<monogr>
		<title level="m">Automated Reasoning -11th International Joint Conference, IJCAR 2022</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">J</forename><surname>Blanchette</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Kovács</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Pattinson</surname></persName>
		</editor>
		<meeting><address><addrLine>Haifa, Israel</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2022">August 8-10, 2022. 2022</date>
			<biblScope unit="volume">13385</biblScope>
			<biblScope unit="page" from="659" to="667" />
		</imprint>
	</monogr>
	<note>Proceedings</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Boldly going where no prover has gone before</title>
		<author>
			<persName><forename type="first">G</forename><surname>Reger</surname></persName>
		</author>
		<idno type="DOI">10.4204/EPTCS.311.6</idno>
		<ptr target="https://doi.org/10.4204/EPTCS.311.6.doi:10.4204/EPTCS.311.6" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Second International Workshop on Automated Reasoning: Challenges, Applications, Directions, Exemplary Achievements, ARCADE@CADE 2019</title>
				<editor>
			<persName><forename type="first">M</forename><surname>Suda</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Winkler</surname></persName>
		</editor>
		<meeting>the Second International Workshop on Automated Reasoning: Challenges, Applications, Directions, Exemplary Achievements, ARCADE@CADE 2019<address><addrLine>Natal, Brazil</addrLine></address></meeting>
		<imprint>
			<publisher>EPTCS</publisher>
			<date type="published" when="2019-08-26">August 26, 2019. 2019</date>
			<biblScope unit="volume">311</biblScope>
			<biblScope unit="page" from="37" to="41" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Towards efficient subsumption</title>
		<author>
			<persName><forename type="first">T</forename><surname>Tammet</surname></persName>
		</author>
		<idno type="DOI">10.1007/BFb0054276</idno>
		<ptr target="https://doi.org/10.1007/BFb0054276.doi:10.1007/BFb0054276" />
	</analytic>
	<monogr>
		<title level="m">Automated Deduction -CADE-15, 15th International Conference on Automated Deduction</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">C</forename><surname>Kirchner</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Kirchner</surname></persName>
		</editor>
		<meeting><address><addrLine>Lindau, Germany</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="1998">July 5-10, 1998. 1998</date>
			<biblScope unit="volume">1421</biblScope>
			<biblScope unit="page" from="427" to="441" />
		</imprint>
	</monogr>
	<note>Proceedings</note>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m" type="main">Regularization in Spider-style strategy discovery and schedule construction</title>
		<author>
			<persName><forename type="first">F</forename><surname>Bártek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Chvalovský</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Suda</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2403.12869</idno>
		<idno type="arXiv">arXiv:2403.12869</idno>
		<ptr target="https://arxiv.org/abs/2403.12869.doi:10.48550/arXiv.2403.12869" />
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Heterogeneous heuristic optimisation and scheduling for first-order theorem proving</title>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">K</forename><surname>Holden</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Korovin</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-030-81097-9_8</idno>
		<ptr target="https://doi.org/10.1007/978-3-030-81097-9_8.doi:10.1007/978-3-030-81097-9\_8" />
	</analytic>
	<monogr>
		<title level="m">Intelligent Computer Mathematics -14th International Conference, CICM 2021</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">F</forename><surname>Kamareddine</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><forename type="middle">S</forename><surname>Coen</surname></persName>
		</editor>
		<meeting><address><addrLine>Timisoara, Romania</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2021">July 26-31, 2021. 2021</date>
			<biblScope unit="volume">12833</biblScope>
			<biblScope unit="page" from="107" to="123" />
		</imprint>
	</monogr>
	<note>Proceedings</note>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<title level="m" type="main">Classification and Regression Trees</title>
		<author>
			<persName><forename type="first">L</forename><surname>Breiman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">H</forename><surname>Friedman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">A</forename><surname>Olshen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">J</forename><surname>Stone</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1984">1984</date>
			<pubPlace>Wadsworth</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">First-order theorem proving and Vampire</title>
		<author>
			<persName><forename type="first">L</forename><surname>Kovács</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Voronkov</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-642-39799-8_1</idno>
		<idno>doi:</idno>
		<ptr target="10.1007/978-3-642-39799-8\_1" />
	</analytic>
	<monogr>
		<title level="m">Computer Aided Verification -25th International Conference, CAV 2013</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">N</forename><surname>Sharygina</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Veith</surname></persName>
		</editor>
		<meeting><address><addrLine>Saint Petersburg, Russia</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2013">July 13-19, 2013. 2013</date>
			<biblScope unit="volume">8044</biblScope>
			<biblScope unit="page" from="1" to="35" />
		</imprint>
	</monogr>
	<note>Proceedings</note>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">The CADE ATP System Competition -CASC</title>
		<author>
			<persName><forename type="first">G</forename><surname>Sutcliffe</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">AI Magazine</title>
		<imprint>
			<biblScope unit="volume">37</biblScope>
			<biblScope unit="page" from="99" to="101" />
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<title level="m" type="main">The CADE ATP system competition</title>
		<author>
			<persName><forename type="first">G</forename><surname>Sutcliffe</surname></persName>
		</author>
		<ptr target="https://tptp.org/CASC/" />
		<imprint>
			<date type="published" when="2024-04-26">2024. 26-April-2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">Spider: Learning in the sea of options</title>
		<author>
			<persName><forename type="first">A</forename><surname>Voronkov</surname></persName>
		</author>
		<ptr target="https://easychair.org/smart-program/Vampire23/2023-07-05.html#talk:223833" />
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note>unpublished. Paper accepted at Vam-pire23: The 7th Vampire Workshop</note>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">The TPTP problem library and associated infrastructure</title>
		<author>
			<persName><forename type="first">G</forename><surname>Sutcliffe</surname></persName>
		</author>
		<idno type="DOI">10.1007/s10817-017-9407-7</idno>
	</analytic>
	<monogr>
		<title level="j">Journal of Automated Reasoning</title>
		<imprint>
			<biblScope unit="volume">59</biblScope>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">Vampire strategy performance measurements</title>
		<author>
			<persName><forename type="first">F</forename><surname>Bártek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Suda</surname></persName>
		</author>
		<idno type="DOI">10.5281/zenodo.10814478</idno>
		<ptr target="https://doi.org/10.5281/zenodo.10814478.doi:10.5281/zenodo.10814478" />
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<author>
			<persName><forename type="first">G</forename><surname>Sutcliffe</surname></persName>
		</author>
		<ptr target="https://tptp.org/CASC/J12/Design.html" />
		<title level="m">CASC design and organization</title>
				<imprint>
			<date type="published" when="2024-04-15">2024. 15-April-2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Greedy function approximation: A gradient boosting machine</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">H</forename><surname>Friedman</surname></persName>
		</author>
		<idno type="DOI">10.1214/aos/1013203451</idno>
		<ptr target="https://doi.org/10.1214/aos/1013203451.doi:10.1214/aos/1013203451" />
	</analytic>
	<monogr>
		<title level="j">The Annals of Statistics</title>
		<imprint>
			<biblScope unit="volume">29</biblScope>
			<biblScope unit="page" from="1189" to="1232" />
			<date type="published" when="2001">2001</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">LightGBM: A highly efficient gradient boosting decision tree</title>
		<author>
			<persName><forename type="first">G</forename><surname>Ke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Meng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Finley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Ye</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T.-Y</forename><surname>Liu</surname></persName>
		</author>
		<ptr target="https://proceedings.neurips.cc/paper_files/paper/2017/file/6449f44a102fde848669bdd9eb6b76fa-Paper.pdf" />
	</analytic>
	<monogr>
		<title level="m">Advances in Neural Information Processing Systems</title>
				<editor>
			<persName><forename type="first">I</forename><surname>Guyon</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">U</forename><forename type="middle">V</forename><surname>Luxburg</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Bengio</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Wallach</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Fergus</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Vishwanathan</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Garnett</surname></persName>
		</editor>
		<imprint>
			<publisher>Curran Associates, Inc</publisher>
			<date type="published" when="2017">2017</date>
			<biblScope unit="volume">30</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">The algorithm selection problem</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">R</forename><surname>Rice</surname></persName>
		</author>
		<idno type="DOI">10.1016/S0065-2458(08)60520-3</idno>
		<idno>doi:10.1016/S0065-2458(08)60520-3</idno>
		<ptr target="https://doi.org/10.1016/S0065-2458(08)60520-3" />
	</analytic>
	<monogr>
		<title level="j">Adv. Comput</title>
		<imprint>
			<biblScope unit="volume">15</biblScope>
			<biblScope unit="page" from="65" to="118" />
			<date type="published" when="1976">1976</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Automated algorithm selection: Survey and perspectives</title>
		<author>
			<persName><forename type="first">P</forename><surname>Kerschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">H</forename><surname>Hoos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Neumann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Trautmann</surname></persName>
		</author>
		<idno type="DOI">10.1162/EVCO_A_00242</idno>
		<ptr target="https://doi.org/10.1162/evco_a_00242.doi:10.1162/EVCO\_A\_00242" />
	</analytic>
	<monogr>
		<title level="j">Evol. Comput</title>
		<imprint>
			<biblScope unit="volume">27</biblScope>
			<biblScope unit="page" from="3" to="45" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<author>
			<persName><forename type="first">S</forename><surname>Schulz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Cruanes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Vukmirovic</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-030-29436-6_29</idno>
		<idno>doi:</idno>
		<ptr target="10.1007/978-3-030-29436-6\_29" />
	</analytic>
	<monogr>
		<title level="m">Automated Deduction -CADE 27 -27th International Conference on Automated Deduction</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">P</forename><surname>Fontaine</surname></persName>
		</editor>
		<meeting><address><addrLine>Natal, Brazil</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2019">August 27-30, 2019. 2019</date>
			<biblScope unit="volume">11716</biblScope>
			<biblScope unit="page" from="495" to="507" />
		</imprint>
	</monogr>
	<note>Proceedings</note>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">MaLeS: A framework for automatic tuning of automated theorem provers</title>
		<author>
			<persName><forename type="first">D</forename><surname>Kühlwein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Urban</surname></persName>
		</author>
		<idno type="DOI">10.1007/s10817-015-9329-1</idno>
		<ptr target="https://doi.org/10.1007/s10817-015-9329-1.doi:10.1007/s10817-015-9329-1" />
	</analytic>
	<monogr>
		<title level="j">J. Autom. Reason</title>
		<imprint>
			<biblScope unit="volume">55</biblScope>
			<biblScope unit="page" from="91" to="116" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Machine learning for first-order theorem provinglearning to select a good heuristic</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">P</forename><surname>Bridge</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">B</forename><surname>Holden</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">C</forename><surname>Paulson</surname></persName>
		</author>
		<idno type="DOI">10.1007/S10817-014-9301-5</idno>
		<ptr target="https://doi.org/10.1007/s10817-014-9301-5.doi:10.1007/S10817-014-9301-5" />
	</analytic>
	<monogr>
		<title level="j">J. Autom. Reason</title>
		<imprint>
			<biblScope unit="volume">53</biblScope>
			<biblScope unit="page" from="141" to="172" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Targeted configuration of an SMT solver</title>
		<author>
			<persName><forename type="first">J</forename><surname>Hůla</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Jakubův</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Janota</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Kubej</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-031-16681-5_18</idno>
		<idno>doi:</idno>
		<ptr target="10.1007/978-3-031-16681-5\_18" />
	</analytic>
	<monogr>
		<title level="m">Intelligent Computer Mathematics -15th International Conference, CICM 2022</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">K</forename><surname>Buzzard</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><surname>Kutsia</surname></persName>
		</editor>
		<meeting><address><addrLine>Tbilisi, Georgia</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2022">September 19-23, 2022. 2022</date>
			<biblScope unit="volume">13467</biblScope>
			<biblScope unit="page" from="256" to="271" />
		</imprint>
	</monogr>
	<note>Proceedings</note>
</biblStruct>

<biblStruct xml:id="b21">
	<monogr>
		<title/>
		<author>
			<persName><forename type="first">S</forename><surname>Schulz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E-</forename><surname>Setheo</surname></persName>
		</author>
		<ptr target="http://wwwlehre.dhbw-stuttgart.de/~sschulz/WORK/e-setheo.html" />
		<imprint>
			<date type="published" when="2024-04-26">2024. 26-April-2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Case-based reasoning for autonomous constraint solving</title>
		<author>
			<persName><forename type="first">D</forename><surname>Bridge</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>O'mahony</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>O'sullivan</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-642-21434-9_4</idno>
		<ptr target="https://doi.org/10.1007/978-3-642-21434-9_4.doi:10.1007/978-3-642-21434-9\_4" />
	</analytic>
	<monogr>
		<title level="m">Autonomous Search</title>
				<editor>
			<persName><forename type="first">Y</forename><surname>Hamadi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">É</forename><surname>Monfroy</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Saubion</surname></persName>
		</editor>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2012">2012</date>
			<biblScope unit="page" from="73" to="95" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">BliStrTune: hierarchical invention of theorem proving strategies</title>
		<author>
			<persName><forename type="first">J</forename><surname>Jakubův</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Urban</surname></persName>
		</author>
		<idno type="DOI">10.1145/3018610.3018619</idno>
		<idno>doi:10.1145/3018610.3018619</idno>
		<ptr target="https://doi.org/10.1145/3018610.3018619" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 6th ACM SIGPLAN Conference on Certified Programs and Proofs, CPP 2017</title>
				<editor>
			<persName><forename type="first">Y</forename><surname>Bertot</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">V</forename><surname>Vafeiadis</surname></persName>
		</editor>
		<meeting>the 6th ACM SIGPLAN Conference on Certified Programs and Proofs, CPP 2017<address><addrLine>Paris, France</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2017">January 16-17, 2017. 2017</date>
			<biblScope unit="page" from="43" to="52" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
