<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Two ways of using artificial neural networks in knowledge discovery from chemical materials data ⋆</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Martin</forename><surname>Holeňa</surname></persName>
							<email>martin@cs.cas.cz</email>
							<affiliation key="aff0">
								<orgName type="department">Institute of Computer Science</orgName>
								<orgName type="institution">Academy of Sciences of the Czech Republic Pod Vodárenskou věží</orgName>
								<address>
									<postCode>18207</postCode>
									<settlement>Prague</settlement>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Two ways of using artificial neural networks in knowledge discovery from chemical materials data ⋆</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">68E26E141FC7A9A36C3337D6FDB1839F</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T14:30+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In the application area of chemical materials, data mining methods have been used for more than a decade. By far most popular have from the very beginning been methods based on artificial neural networks. However, they are frequently used without awareness of the difference between the numeric nature of knowledge obtained from data by neural network regression, and the symbolic nature of knowledge obtained by some other data mining methods. This paper explains that within the surrogate modelling approach, which plays an important role in this area, using numeric knowledge is justified. At the same time, it recalls the possibility to obtain symbolic knowledge from neural networks in the form of logical rules and describes a recently proposed method for the extraction of Boolean rules in disjunctive normal form. Both ways of using neural networks are illustrated on examples from this application area.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>The search for new chemical materials, e.g., catalytic materials for a plethora of chemical reactions, produces large amounts of data. To discover useful knowledge from those data, statistical as well as machinelearning data mining methods have been used in this area since the late 1990s, the former represented in particular by the analysis of variance, decision trees and support vector regression, the latter by main variants of feed-forward neural networks.</p><p>This paper summarizes experience from nearly ten years using and developing neural-networks based data mining methods for catalytic data. Artificial neural networks are the most popular regression model in this application area. In the survey <ref type="bibr" target="#b9">[10]</ref>, more than 20 published applications of multilayer perceptrons (MLPs) to catalytic data have been listed, as well as several applications of radial basis function networks. The role of feed-forward neural nets as a regression model predicting catalytic performance of materials (such as yield, conversion, selectivity) is due partially to their preceding success in other areas, but mostly to their ability ⋆ The research reported in this paper has been supported by the grant No. 201/08/1744 of the Grant Agency of the Czech Republic and partially supported by the Institutional Research Plan AV0Z10300504.</p><p>to serve as universal approximators in very general function spaces <ref type="bibr" target="#b11">[12,</ref><ref type="bibr" target="#b13">14,</ref><ref type="bibr" target="#b17">18]</ref>. This ability is particularly valuable in the context of the highly nonlinear nature of the dependencies encountered in catalysis (cf. Figure <ref type="figure" target="#fig_0">1</ref>). However, it seems to be little awareness, among researchers using artificial neural networks in catalysis, of the difference among the symbolic nature of the knowledge obtained from data by analysis of variance and decision trees, and the numeric nature of the knowledge obtained by neural network regression. Incited by the situation just outlined, the paper presents two strategies for the application of artificial neural networks to data about chemical materials. The first strategy relies on the numeric knowledge from neural network regression. Although numeric knowledge is much less understandable to humans than symbolic knowledge (in terms of <ref type="bibr" target="#b3">[4]</ref>, it has a high "data fit", but a low "mental fit"), we show that in this application area, it can be very useful if directly integrated with the optimization of materials performance in an approach called surrogate modelling. In that context, also the possibility to increase the accuracy of neural network regression by means of boosting is mentioned. The other strategy, on the other hand, relies on employing rules extraction methods to obtain, from trained neural networks, symbolic knowledge.</p><p>These two strategies determine also the structure of the paper. In Section 2, the surrogate modelling approach is described. Section 3 then explains a method for the extraction of logical rules from trained neural networks. Both strategies are in the respective sections illustrated using real-world examples.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Neural networks used as surrogate models</head><p>From the point of view of theoretical computer science, the search for most suitable chemical materials entails complex optimization tasks. As objective functions, those tasks use various properties of the materials, e.g. in the case of catalytic materials, properties quantifying their catalytic performance, such as yield, conversion, or selectivity. A crucial feature of such objective functions is that they cannot be expressed analytically, their values must be obtained empirically.</p><p>For their optimization, it is not possible to employ most common optimization methods, such as steepest descent, conjugate gradient methods or the Levenberg-Marquardt method. Indeed, to obtain sufficiently precise numerical estimates of gradients or second order derivatives of the empirical objective function, those methods need to evaluate the function in points some of which would have a smaller distance than is the empirical error of catalytic measurements. That is why methods not requiring any derivatives have been used to solve such optimization tasks, such as the simplex method, and most frequently genetic and other evolutionary algorithms <ref type="bibr" target="#b1">[2]</ref>. To compensate for missing information about derivatives, these methods need quite large number of objective function evaluations. In the context of catalysis, this is quite disadvantageous because the evaluation of the empirical objective functions used in the search for optimal catalysts is often costly and time-consuming. Testing a generation of catalytic materials proposed by an evolutionary algorithm typically needs several days of time and costs thousands of euros.</p><p>The usual approach to decreasing the cost and time of optimization of empirical objective functions is to evaluate the function only in points considered to be most important for the progress of the employed optimization method, and to evaluate its suitable regression model otherwise. That model is termed surrogate model of the function, and the approach is referred to as surrogate modelling <ref type="bibr" target="#b16">[17,</ref><ref type="bibr" target="#b19">20,</ref><ref type="bibr" target="#b22">23,</ref><ref type="bibr" target="#b26">27]</ref>. Needless to say, the time and costs needed to evaluate a regression model are negligible compared to time and costs needed to evaluate empirical functions such as yield or conversion. However, it must not be forgotten that the agreement between the results obtained with a surrogate model and those obtained with the original function depends on the accuracy of the model.</p><p>The fact that feed-forward neural networks are the most frequent regression models in catalysis suggests them as the most natural candidate for surrogate models in this area. Indeed, several nice examples of the application of neural-network based surrogate modelling to the optimization of performance of catalytic materials have been published during the last five years <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b20">21,</ref><ref type="bibr" target="#b24">25]</ref>. Within the overall context of the application of artificial neural networks to mining catalytic data, however, they are still rare.</p><p>Although surrogate modelling has been also applied to conventional optimization <ref type="bibr" target="#b4">[5]</ref>, it is most frequently encountered in connection with evolutionary algorithms because for them, the approach leads to the approximation of the fitness function, whose usefulness in evolutionary computation is already known <ref type="bibr" target="#b12">[13,</ref><ref type="bibr" target="#b18">19]</ref>. For the progress of evolutionary optimization, most important criteria are on the one hand points that indicate closeness to the global optimum (through highest values of the fitness function), on the other hand points that most contribute to the diversity of the population.</p><p>In the literature, various possibilities of combining evolutionary optimization with surrogate modelling have been discussed <ref type="bibr" target="#b16">[17,</ref><ref type="bibr" target="#b23">24,</ref><ref type="bibr" target="#b26">27]</ref>. Nevertheless, all of them are controlled by one of two basic approaches:</p><p>A. The individual-based-control consists in choosing between the evaluation of the empirical objective function and the evaluation of its surrogate model individual-wise, basically in the following steps: (i) An initial set E of individuals is collected, in which the considered empirical fitness η was evaluated (for example, the population of several first generations of the evolutionary algorithm). (ii) The surrogate model is constructed using the set of pairs {(x, η(x)) : x ∈ E}. (iii) The evolutionary algorithm is run with the fitness η replaced by the model for one generation with a population Q of size qP , where P is the desired population size for the optimization of η, and q is a prescribed ratio (e.g., q = 10 or q = 100). (iv) A subset P ⊂ Q of size P is selected so as to contain those individuals from Q that are most important according to the considered criteria for the progress of optimization.</p><p>(v) For x ∈ P, the empirical fitness is evaluated.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>(vi)</head><p>The set E is replaced by E ∪ P and the algorithm returns to the step (ii).</p><p>B. The generation-based-control consists in choosing between both kinds of evaluation generation-wise, basically in the following steps: The agreement between the results that are obtained with a surrogate model and those that would be obtained if the empirical objective function were evaluated depends on the accuracy of the model. A popular approach to increasing the accuracy of learning methods is boosting, i.e., construction of a strong learner through combining weak learners. It is important to realize that boosted surrogate models are only particular kinds of surrogate models and their interaction with optimization algorithms in optimization tasks follows the same rules as the interaction of surrogate models in general. In particular in the above outlines of individual-based and generation-based control, boosting is always performed in the step (ii), which has to be replaced with:</p><formula xml:id="formula_0">(i) An initial set E of</formula><formula xml:id="formula_1">(ii'a) The set {(x, η(x)) : x ∈ E} is divided into k disjoint subsets of size ⌊ |E| k ⌋ or ⌈ |E| k ⌉,</formula><p>where | | denotes the cardinality of a set, ⌊ ⌋ the lower integer bound of a real number, and ⌈ ⌉ its upper integer bound.</p><p>(ii'b) For each j = 1, . . . , k, a surrogate model F j 1 is constructed, using only data not belonging to the j-th subset.</p><p>(ii'c) A k-fold crossvalidation of regression boosting is performed, and the error of the boosting approximation is in each iteration measured with the prescribed error measure on the validation data. (ii'd) The first iteration i in which the average error of the boosting approximation on the validation data is lower than in the i + 1-th iteration is taken as the final iteration of boosting. (ii'e) Boosting using the complete set {(x, η(x)) :</p><p>x ∈ E} is performed up to the final iteration found in step (ii'd), and the result of the application of the employed boosting method in each such iteration of boosting is taken as the boosted surrogate model in that iteration.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">An illustration</head><p>A particular method for MLP boosting has been presented in <ref type="bibr" target="#b10">[11]</ref>. That method will now be employed in surrogate modelling with data from the investigation of catalytic materials for the high-temperature synthesis of hydrocyanic acid (HCN) <ref type="bibr" target="#b15">[16]</ref>. The composition of most of those materials was designed by means of a specific genetic algorithm (GA) for heterogeneous catalysis <ref type="bibr" target="#b25">[26]</ref>. As usually in evolutionary optimization of catalytic materials, the GA configuration was determined by the experimental conditions in which the optimization was performed: number of channels of the reactor in which the materials were tested, as well as time and financial resources available for those expensive tests. In the reported investigation, the algorithm was running for 7 generations of population size 92, and in addition 52 other catalysts with manually designed composition were investigated. Consequently, data about 696 catalytic materials were available. The considered MLPs had 14 input neurons: 4 of them coding catalyst support, the other 10 corresponding to the proportions of 10 metal additives forming the active shell, and 3 output neurons, corresponding to 3 kinds of catalytic activity considered as fitness functions.</p><p>For boosting, only data about catalysts from the 1.-6. generation of the GA and about the 52 catalysts with manually designed composition were used, thus altogether data about 604 catalytic materials. Data about catalysts from the 7. generation were completely excluded and left out for testing. The set of architectures to which boosting was applied was restricted to MLPS with 1 and 2 hidden layers and was delimited by means of the heuristic pyramidal condition: the number of neurons in a subsequent layer must not exceed the number of neurons in a previous layer. Let n I , n H and n O denote the numbers of input, hidden and output neurons, respectively, and n H1 and n H2 denote the numbers of neurons in the first and second hid-den layer, respectively. Then the pyramidal condition entails the following 90 architectures: (i) one hidden layer and 3 ≤ n H ≤ 14 (12 architectures); (ii) two hidden layers and 3 ≤ n H2 ≤ n H1 ≤ 14 (78 architectures). As was mentioned above, boosting can be combined both with the individaul-based and with the generation-based control of surrogate modelling. In the reported investigation of catalytic materials for HCN synthesis, the indiviual-based control was employed.</p><p>The error measure employed in the crossvalidation in the step (ii'c) was MSE. The distribution of the final iterations of boosting, found for MLPs with the 90 considered architectures in the step (ii'd), is depicted in Figure <ref type="figure" target="#fig_1">2</ref>. We can see that only for 16 MLPs, already the 1st iteration was the final. For the remaining 74 MLPs, boosting improved the average MSE on the validation data for at least 1 iteration. The mean and median of the distribution of the final iterations were 6.6 and 5, respectively. For testing with the data from the 7th generation of the evolutionary algorithm, we used only the five MLPs most promising from the point of view of the average MSE on the validation data in the final iteration of boosting. These were the following MLPs:</p><p>a 1-hidden-layer MLP, with n H = 11 and the 3rd iteration of boosting being the final iteration, four 2-hidden-layers MLPs, with (n H1 , n H2 ) = = (10, 4), <ref type="bibr" target="#b9">(10,</ref><ref type="bibr" target="#b5">6)</ref>, <ref type="bibr" target="#b12">(13,</ref><ref type="bibr" target="#b4">5)</ref>, <ref type="bibr" target="#b13">(14,</ref><ref type="bibr" target="#b7">8)</ref> and the final iterations of boosting 19, 32, 31 and 29, respectively.</p><p>For each of them, the validation proceeded as follows:</p><p>1. In each iteration up to the final, a single MLP was trained with data about all the 604 catalytic materials used for boosting.</p><p>2. In each iteration up the final iteration of boosting, the boosted surrogate model was constructed for the trained MLP, according to the step (ii'e). 3. From the values predicted by the boosted surrogate model for the 92 materials from the 7. generation of the GA, and from the measured values, the boosting MSE was calculated.</p><p>The results are summarized in Figure <ref type="figure" target="#fig_2">3</ref>, decomposed to the properties corresponding to the MLP outputs -conversions of CH 4 and NH 3 and yield of HCN. They clearly confirm the usefulness of boosting for the five considered architectures. For each of them, boosting leads to an overall decrease of MSE of the conversion of CH 4 and HCN yield, on new data from the 7th generation of the GA, which is uninterrupted or nearly uninterrupted till the final boosting iteration. On the other hand, boosting did not lead to any decrease of the error of the conversion of NH3, which on the other hand is already from the beginning much lower than the two other performance measures (notice that the scale of the y-axis is 10-times finer for the conversion of NH 3 than for the conversion of CH 4 and HCN yield). The explanation for the different behavior of the conversion of NH 3 is the substantially lower variability of its values in the seventh generation of the GA, used for validating the usefulness of boosting (standard deviation, SD: 2.8, interquartile range, IQR: 1.6), compared to the conversion of CH 4 (SD: 26.1, IQR: 45.0) and HCN yield (SD: 20.1, IQR: 35.9). Due to so low variability, the conversion of NH 3 appears effectively as nearly constant during the validation of boosting, which in turn accounts for a nearly constant MSE. In addition, also implications and equivalences of important kinds of fuzzy logic are frequently extracted <ref type="bibr" target="#b7">[8,</ref><ref type="bibr" target="#b14">15]</ref>. In general, extracted formulas of a formal logic are called rules. Over the last two decades, various rules extraction methods have been proposed for neural networks, but so far none of them has become a common standard (cf. the survey papers <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b14">15,</ref><ref type="bibr" target="#b21">22]</ref> and the monograph <ref type="bibr" target="#b6">[7]</ref>). Here, a method for the extraction of Boolean implications from multilayer perceptrons with n inputs and m outputs will be sketched that finds to each output condition of the form:</p><formula xml:id="formula_2">C O : the value y of the output variables lies in a rectangular area R ⊂ R m (2)</formula><p>one or more input conditions of the form C I : the value x of the input variables lies in a polyhedron P ⊂ R n (3)</p><p>Hence, this method extracts rules of the form:</p><formula xml:id="formula_3">IF x ∈ P THEN y ∈ R.<label>(4)</label></formula><p>A detailed explanation of the method can be found in <ref type="bibr" target="#b8">[9]</ref>. Its main principles can be summarized as follows:</p><p>-An m-dimensional rectangular area R with borders perpendicular to the m coordinate axes has to be chosen in advance in the output space of a trained MLP with sigmoid activation functions. The reason for choosing such an area is that in the space of evaluations of m free variables, each m-dimensional rectangular area is the validity set of the conjunction of some m univariate Boolean predicates. That conjunction then serves as the consequent of the rule to extract. -The activation functions in the hidden neurons are approximated with piecewise-linear sigmoid activation functions. This can be done with an arbitrary precision. -The products of individual linearity intervals of all the activation functions determine areas in the input space in which the final approximating mapping computed by the multilayer perceptron is linear.</p><p>-In each such area, all points mapped to R form a polyhedron, which may eventually be empty or may be concatenated with polyhedra from some of the neighboring areas to a larger polyhedron. -The union of all the nonempty concatenated polyhedra P 1 , . . . , P q defines the antecedent of a rule in a combined form</p><formula xml:id="formula_4">IF x ∈ P 1 ∪ • • • ∪ P q THEN y ∈ R,<label>(5)</label></formula><p>which is equivalent to a logical disjunction of q rules of the simple form (4):</p><formula xml:id="formula_5">IF x ∈ P 1 THEN y ∈ R . . . (6) IF x ∈ P q THEN y ∈ R.</formula><p>To increase the comprehensibility of the extracted rules, visualization by means of 2-or 3-dimensional cuts of the set P 1 ∪ • • • ∪ P q can be used (Figure <ref type="figure" target="#fig_3">4</ref>).</p><p>Usually, logical rules of the form (4) are the final results of this rule-extraction method. Nonetheless, there is one exception -when the polyhedron P is also rectangular with borders perpendicular to axes, or more generally, when P can be approximately replaced with such a rectangular area R I in the input space. Then the above rule (4) can be approximately expressed in the conjunctive form</p><formula xml:id="formula_6">IF x 1 ∈ I 1 &amp; . . . &amp; x nI ∈ I nI THEN y ∈ R. (<label>7</label></formula><formula xml:id="formula_7">)</formula><p>Here, I 1 , . . . , I nI are intervals that constitute the projections of R I into the n I input dimensions. Each such interval can be restricted both from below and from above, restricted only from below or only from above, or finally can be even the complete set of real numbers. However, dimensions for which the corresponding projection of R I equals the complete real axis are usually not included in <ref type="bibr" target="#b6">(7)</ref>, since they would not provide any new knowledge. Finally, observe that due to ( <ref type="formula" target="#formula_4">5</ref>) and ( <ref type="formula" target="#formula_6">7</ref>), the final extracted rule is in the disjunctive normal form.</p><p>In the rule-extraction method outlined above, the possibility of replacing a polyhedron P with a rectangular area R I is assessed according to the following principles:</p><p>1. The resulting dissatisfaction with points that either belong to P but do not belong to R I , or belong to R I but do not belong to P (i.e., with points from the symmetric difference R I ∆P ), has to remain within a prescribed tolerance ε and R I has to be minimal in the input space among rectangular areas of some specified kind with dissatisfacion within that tolerance. 2. The dissatisfaction with points from R I ∆P depends solely on those points and is increasing with respect to inclusion. Consequently, it can be measured using some monotone measure on the input space, possibly depending on P . 3. To be eligible for replacement, P has to cover at least one point of the available data.</p><p>For 2., the most attractive monotone measures, due to their straightforward interpretability, are:</p><p>-The joint empirical distribution of the input variables in the available data.</p><p>-The conditional empirical distribution of the input variables in the available data, conditioned by P .</p><p>Rules of the form (7) are also very convenient from the visualization point of view: Since cuts of rectangular areas coincide with the corresponding projections of those areas, the values of no variables need to be fixed.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">An illustration</head><p>As an example, Figure <ref type="figure">5</ref> shows three-dimensional cuts determining the antecedents of conjunctive-form rules extracted from a trained MLP with 5 input neurons and 1 output neuron such that: (i) the input neurons correspond to variables that record the molar proportions of the oxides of Fe, Ga, Mg, Mn and Mo in the catalytic material; (ii) the output neuron corresponds to a variable recording propene yield. The extracted rules are listed in Table <ref type="table">1</ref>. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Conclusion</head><p>The paper dealt with employing feed-forward neural networks for knowledge discovery from data about chemical materials. It has shown that in this application area, obtaining numeric knowledge by neural-network regression is justified, in spite of the fact that numeric knowledge is substantially less human-understandable Table <ref type="table">1</ref>. Antecedents of the rules of the form (7) extracted using the method described in this section for the consequent "propene yield &gt; 8%" from a trained MLP with 5 input neurons and 1 output neuron, assuming that the above interpretation of the variables to which those neurons correspond is described by (i) and (ii).</p><p>Fig. <ref type="figure">5</ref>. A three-dimensional projection of the union of rectangular areas that replace, following the method described in this section, the union of of polyhedra from the antecedent of a combined form rule extracted from a trained MLP. The projection corresponds to input variables recording the molar proportion of oxides of Ga, Mg and Mo in a catalytic material. The numbers 1, 2, 3 refer to the antecedents of the rules in Table <ref type="table">1</ref>.</p><p>than symbolic knowledge. Its justification consists in the possibility to use such knowledge in the optimization tasks entailed by search for new materials in the surrogate modelling approach.</p><p>In addition to justifying the specific need for numeric knowledge from neural network regression in this application area, the paper recalled the possibility to obtain symbolic knowledge in the form of logical rules from trained neural networks. It explained a recently proposed method for the extraction of Boolean rules in disjunctive normal form, and illustrated it on data about catalytic materials.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. A 3-dimensional cut of a neural-network regression of the yield of a reaction product on the composition of the catalytic material.</figDesc><graphic coords="1,305.13,364.49,233.09,221.12" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. Distribution of the final iterations of boosting of the 90 MLPs with 1-hidden-layer architectures fulfilling 3 ≤ nH ≤ 14 and 2-hidden-layer architectures fulfilling 3 ≤ nH2 ≤ nH1 ≤ 14.</figDesc><graphic coords="4,62.70,357.52,231.61,137.25" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>3Fig. 3 .</head><label>3</label><figDesc>Fig. 3. History of the boosting MSE on the data from the 7th generation of the GA for MLPs with the 5 architectures included in the validation of boosting, decomposed to the properties corresponding to the MLP outputs.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Fig. 4 .</head><label>4</label><figDesc>Fig. 4.A two-dimensional cut of the union of polyhedra from the antecedent of a rule of the form (5) extracted from a trained MLP. The cut corresponds to input variables recording the molar proportions of oxides of Mn and Ga in the catalytic material, for the consequent "propene yield &gt; 8%".</figDesc><graphic coords="6,307.79,363.40,230.35,172.76" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>&lt;</head><label></label><figDesc>Ga proportion &lt; 33% &amp; 31% &lt; Mg proportion &lt; 39% &amp; Mo proportion &lt; 7% &amp; Fe, Mn proportions = 0 2 Ga proportion ≈ 36% &amp; 28% &lt; Mg proportion &lt; 38% C3H6 yield &gt; 8% &amp; Fe, Mn, Mo proportions = 0 3 Fe proportion &lt; 12% &amp; Ga proportion ≈ 38% &amp; 29% &lt; Mg proportion &lt; 36% &amp; Mo proportion &lt; 9% &amp; Mn proportion = 0</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0"><head></head><label></label><figDesc></figDesc><graphic coords="5,62.96,91.01,474.16,261.75" type="bitmap" /></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Survey and critique of techniques for extracting rules from trained artificical neural networks</title>
		<author>
			<persName><forename type="first">R</forename><surname>Andrews</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Diederich</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">B</forename><surname>Tickle</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Knowledge Based Systems</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="page" from="378" to="389" />
			<date type="published" when="1995">1995</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Combinatorial development of solid catalytic materials. Design of high-throughput experiments, data analysis, data mining</title>
		<author>
			<persName><forename type="first">M</forename><surname>Baerns</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Holeňa</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="s">World Scientific</title>
		<imprint>
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Using artificial neural networks to boost high-throughput discovery in heterogeneous catalysis</title>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">A</forename><surname>Baumes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Farrusseng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Lengliz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Mirodatos</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">QSAR and Combinatorial Science</title>
		<imprint>
			<biblScope unit="volume">23</biblScope>
			<biblScope unit="page" from="767" to="778" />
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m">Intelligent data analysis. An introduction</title>
				<editor>
			<persName><forename type="first">M</forename><surname>Berthold</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Hand</surname></persName>
		</editor>
		<meeting><address><addrLine>Berlin</addrLine></address></meeting>
		<imprint>
			<publisher>Springer Verlag</publisher>
			<date type="published" when="1999">1999</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Trosset: A rigorous framework for optimization by surrogates</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">J</forename><surname>Brooker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Dennis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">D</forename><surname>Frank</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">B</forename><surname>Serafini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Torczon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Structural and Multidisciplinary Optimization</title>
		<imprint>
			<biblScope unit="volume">17</biblScope>
			<biblScope unit="page" from="1" to="13" />
			<date type="published" when="1998">1998</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Development of an integrated informatics toolbox: HT kinetic and virtual screening</title>
		<author>
			<persName><forename type="first">D</forename><surname>Farrusseng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Clerc</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Mirodatos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Azam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Gilardoni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">W</forename><surname>Thybaut</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Balasubramaniam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">B</forename><surname>Marin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Combinatorial Chemistry and High Throughput Screening</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="page" from="85" to="97" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m" type="main">Neuralsymbolic cognitive reasoning</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">S A</forename><surname>Garcez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">C</forename><surname>Lamb</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">M</forename><surname>Gabbay</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2009">2009</date>
			<publisher>Springer Verlag</publisher>
			<pubPlace>Berlin</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Extraction of fuzzy logic rules from data by means of artificial neural networks</title>
		<author>
			<persName><forename type="first">M</forename><surname>Holeňa</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Kybernetika</title>
		<imprint>
			<biblScope unit="volume">41</biblScope>
			<biblScope unit="page" from="297" to="314" />
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Piecewise-linear neural networks and their relationship to rule extraction from data</title>
		<author>
			<persName><forename type="first">M</forename><surname>Holeňa</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Neural Computation</title>
		<imprint>
			<biblScope unit="volume">18</biblScope>
			<biblScope unit="page" from="2813" to="2853" />
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Computer-aided strategies for catalyst development</title>
		<author>
			<persName><forename type="first">M</forename><surname>Holeňa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Baerns</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Handbook of Heterogeneous Catalysis</title>
				<editor>
			<persName><forename type="first">G</forename><surname>Ertl</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Knözinger</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Schüth</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Eitkamp</surname></persName>
		</editor>
		<meeting><address><addrLine>Weinheim</addrLine></address></meeting>
		<imprint>
			<publisher>Wiley-VCH</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="66" to="81" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Boosted neural networks in evolutionary computation</title>
		<author>
			<persName><forename type="first">M</forename><surname>Holeňa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Linke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Steinfeldt</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Neural Information Processing</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<meeting><address><addrLine>Berlin</addrLine></address></meeting>
		<imprint>
			<publisher>Springer Verlag</publisher>
			<date type="published" when="2009">2009</date>
			<biblScope unit="volume">5864</biblScope>
			<biblScope unit="page" from="131" to="140" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Approximation capabilities of multilayer neural networks</title>
		<author>
			<persName><forename type="first">K</forename><surname>Hornik</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Neural Networks</title>
		<imprint>
			<biblScope unit="volume">4</biblScope>
			<biblScope unit="page" from="251" to="257" />
			<date type="published" when="1991">1991</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Sendhoff: Neural networks for fitness approximation in evolutionary optimization</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Jin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hüsken</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Olhofer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Knowledge Incorporation in Evolutionary Computation</title>
				<editor>
			<persName><forename type="first">Y</forename><surname>Jin</surname></persName>
		</editor>
		<meeting><address><addrLine>Berlin</addrLine></address></meeting>
		<imprint>
			<publisher>Springer Verlag</publisher>
			<date type="published" when="2005">2005</date>
			<biblScope unit="page" from="281" to="306" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Neural networks as universal approximators</title>
		<author>
			<persName><forename type="first">V</forename><surname>Kůrková</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Handbook of Brain Theory and Neural Networks</title>
				<editor>
			<persName><forename type="first">M</forename><surname>Arbib</surname></persName>
		</editor>
		<meeting><address><addrLine>Cambridge</addrLine></address></meeting>
		<imprint>
			<publisher>MIT Press</publisher>
			<date type="published" when="2002">2002</date>
			<biblScope unit="page" from="1180" to="1183" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Neuro-fuzzy rule generation: Survey in soft computing framework</title>
		<author>
			<persName><forename type="first">S</forename><surname>Mitra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Hayashi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Neural Networks</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<biblScope unit="page" from="748" to="768" />
			<date type="published" when="2000">2000</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">New catalytic materials for the high-temperature synthesis of hydrocyanic acid from methane and ammonia by high-throughput approach</title>
		<author>
			<persName><forename type="first">S</forename><surname>Möhmel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Steinfeldt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Endgelschalt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Holeňa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kolf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Dingerdissen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Wolf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Weber</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Bewersdorf</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Applied Catalysis A: General</title>
		<imprint>
			<biblScope unit="volume">334</biblScope>
			<biblScope unit="page" from="73" to="83" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Surrogate-assisted evolutionary optimization frameworks for high-fidelity engineering design problems</title>
		<author>
			<persName><forename type="first">Y</forename><forename type="middle">S</forename><surname>Ong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">B</forename><surname>Nair</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">J</forename><surname>Keane</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">W</forename><surname>Wong</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Knowledge Incorporation in Evolutionary Computation</title>
				<editor>
			<persName><forename type="first">Y</forename><surname>Jin</surname></persName>
		</editor>
		<meeting><address><addrLine>Berlin</addrLine></address></meeting>
		<imprint>
			<publisher>Springer Verlag</publisher>
			<date type="published" when="2005">2005</date>
			<biblScope unit="page" from="307" to="331" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Approximation theory of the MPL model in neural networks</title>
		<author>
			<persName><forename type="first">A</forename><surname>Pinkus</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Acta Numerica</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="page" from="277" to="283" />
			<date type="published" when="1998">1998</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Accelerating the convergence of evolutionary algorithms by fitness landscape approximation</title>
		<author>
			<persName><forename type="first">A</forename><surname>Ratle</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Parallel Problem Solving from Nature</title>
				<editor>
			<persName><forename type="first">A</forename><forename type="middle">E</forename><surname>Eiben</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><surname>Bäck</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Schoenauer</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H.-P</forename><surname>Schwefel</surname></persName>
		</editor>
		<meeting><address><addrLine>Berlin</addrLine></address></meeting>
		<imprint>
			<publisher>Springer Verlag</publisher>
			<date type="published" when="1998">1998</date>
			<biblScope unit="page" from="87" to="96" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Kriging as a surrogate fitness landscape in evolutionary optimization</title>
		<author>
			<persName><forename type="first">A</forename><surname>Ratle</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Artificial Intelligence for Engineering Design, Analysis and Manufacturing</title>
		<imprint>
			<biblScope unit="volume">15</biblScope>
			<biblScope unit="page" from="37" to="49" />
			<date type="published" when="2001">2001</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Application of a genetic algorithm and a neural network for the discovery and optimization of new solid catalytic materials</title>
		<author>
			<persName><forename type="first">U</forename><surname>Rodemerck</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Baerns</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Holeňa</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Applied Surface Science</title>
		<imprint>
			<biblScope unit="volume">223</biblScope>
			<biblScope unit="page" from="168" to="174" />
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">The truth will come to light: Directions and challenges in extracting rules from trained artificial neural networks</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">B</forename><surname>Tickle</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Andrews</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Golea</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Diederich</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Neural Networks</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="page" from="1057" to="1068" />
			<date type="published" when="1998">1998</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Model-assisted steady state evolution strategies</title>
		<author>
			<persName><forename type="first">H</forename><surname>Ulmer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Streichert</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Zell</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">GECCO 2003: Genetic and Evolutionary Computation</title>
				<meeting><address><addrLine>Berlin</addrLine></address></meeting>
		<imprint>
			<publisher>Springer Verlag</publisher>
			<date type="published" when="2003">2003</date>
			<biblScope unit="page" from="610" to="621" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Model assisted evolution strategies</title>
		<author>
			<persName><forename type="first">H</forename><surname>Ulmer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Streichert</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Zell</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Knowledge Incorporation in Evolutionary Computation</title>
				<editor>
			<persName><forename type="first">Y</forename><surname>Jin</surname></persName>
		</editor>
		<meeting><address><addrLine>Berlin</addrLine></address></meeting>
		<imprint>
			<publisher>Springer Verlag</publisher>
			<date type="published" when="2005">2005</date>
			<biblScope unit="page" from="333" to="355" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">DoE framework for catalyst development based on soft computing techniques</title>
		<author>
			<persName><forename type="first">S</forename><surname>Valero</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Argente</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Botti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Serra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Serna</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Moliner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Corma</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computers and Chemical Engineering</title>
		<imprint>
			<biblScope unit="volume">33</biblScope>
			<biblScope unit="page" from="225" to="238" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">An evolutionary approach in the combinatorial selection and optimization of catalytic materials</title>
		<author>
			<persName><forename type="first">D</forename><surname>Wolf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><forename type="middle">V</forename><surname>Buyevskaya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Baerns</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Applied Catalyst A: General</title>
		<imprint>
			<biblScope unit="volume">200</biblScope>
			<biblScope unit="page" from="63" to="77" />
			<date type="published" when="2000">2000</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Combining global and local surrogate models to accellerate evolutionary optimization</title>
		<author>
			<persName><forename type="first">Z</forename><forename type="middle">Z</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><forename type="middle">S</forename><surname>Ong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">B</forename><surname>Nair</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">J</forename><surname>Keane</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">Y</forename><surname>Lum</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Systems, Man and Cybernetics. Part C: Applications and Reviews</title>
		<imprint>
			<biblScope unit="volume">37</biblScope>
			<biblScope unit="page" from="66" to="76" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
