<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Inductive Models of User Preferences for Semantic Web</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Alan</forename><surname>Eckhardt</surname></persName>
							<email>alan.eckhardt@mff.cuni.cz</email>
							<affiliation key="aff0">
								<orgName type="department">Faculty of Mathematics and Physics</orgName>
								<orgName type="institution">Charles University</orgName>
								<address>
									<addrLine>Malostranské nám. 25</addrLine>
									<postCode>118 00</postCode>
									<settlement>Praha 1</settlement>
									<country key="CZ">Czech Republic</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="institution">Charles University in</orgName>
								<address>
									<settlement>Prague</settlement>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Inductive Models of User Preferences for Semantic Web</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">2F3F6A618BFECEDD34FF693B9FB77E5F</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T11:10+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>User preferences became recently a hot topic. The massive use of internet shops and social webs require the presence of a user modelling, which helps users to orient them selfs on a page. There are many different approaches to model user preferences. In this paper, we will overview the current state-of-the-art in the area of acquisition of user preferences and their induction. Main focus will be on the models of user preferences and on the induction of these models, but also the process of extracting preferences from the user behaviour will be studied. We will also present our contribution to the probabilistic user models.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>The user preference modelling plays an important role in the current web. Internet shops need to help the user to find the product he/she searches for, social webs may suggest a contact that the user wants. The process of acquisition of user's preferences starts with the acquisition of known preferences (e.g. from the user behaviour) and then using these known preferences to get the user's preferences of other objects. In this paper, an example of a user who is buying a digital camera will be used. In Section 2, several user models will be presented and in Section 3, some of current methods of induction of user preferences will be described, including our own probabilistic model.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.1">A use case for the induction of user preferences</head><p>We will present a typical use case for the induction of user preferences. We will describe a complex system for the extraction of information from the web and for the presentation of collected information to the user. This system will be aimed to help the user to find a camera that fits best his needs. Whole system proposition is in Figure <ref type="figure" target="#fig_0">1</ref>.</p><p>The first task of this system is to collect data from various sources from the internet. Information is stored in various forms, most often in HTML format, and it has to be transformed to a computer-readable form (Semantic data). The typical computer-readable form is RDF <ref type="bibr" target="#b1">[2]</ref> -a language of triples of the form (subject, predicate, object). Extension of RDF is OWL <ref type="bibr" target="#b0">[1]</ref> which is one of standard ontology languages. Ontologies can be also used to annotate raw text data from the web. This transformation is called the 'semantic annotation' and there are many methods performing semantic annotation, though their accuracy and universality may be questioned. However, studying the semantic annotation is not the aim of this work, it is one of components in whole process.</p><p>With semantically annotated data or with online RDF sources, we may present the integrated information to the user. Several works (e.g. <ref type="bibr" target="#b7">[8]</ref>) concern the graphical user interface that represents the semantic data. The task is to present the most important objects to the user in a such way that he/she will notice them before noticing the other objects.</p><p>The inductive methods enable to determine which objects are important and which are not. The process of determining the importance of an object is iteratively executed, it may be triggered e.g. by some user behaviour, for example by clicking on some object that is not considered important. Interpreting user behaviour is another part of whole process.</p><p>The interpreted behaviour is then used to generate a user preference model. This can be done by an already known inductive method. This user preference model will be then used to alter the appearance of the web page, for presenting preferred objects etc. This information can be also used by other servers, in case of a distributed computation.</p><p>In our work, we will focus on the induction process -the induction of preference model from some rated objects. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Models of user preferences</head><p>We will use the following notation -u 1 , ..., u k ∈ U for the users, o 1 , ..., o n ∈ O for the objects, a 1  1 , ..., a m 1 for the attribute values of the object o 1 and A 1 , ..., A m ∈ A for the attributes of objects. When speaking only of one object, we will also use only o for object and a 1 , ..., a m for its attribute values. Often, all objects are from the same domain and they will have the same set of attributes. For that reason we will be interested mainly in the attribute values. We will denote the user's preference of an object as P (o), meaning how much object o is preferred. We will use the notation P 1 (o), which means the preference of the user u 1 on the object o, when we want to explicitly denote the user. The range of P depends on user model used, some of them do not have direct preference, e.g. the preference relations studied in Section 2.2.</p><p>We now briefly distinguish several types of attribute domains, as was done in <ref type="bibr" target="#b4">[5]</ref>. The first domain type is nominal. These domains have no ordering and are mainly text based. Typical example may be the color or the manufacturer. Second type are ordinal domains, on which exists some ordering, but not unit. For example set Big, Medium, Small may form an ordinal domain. When a unit of measure is added, we get interval domains -for example {1, 0.2, 0, -0.4, -1}. Finally, ratio scales have also an element 0 explicitly defined. Example of ratio scale may be the price, the number of megapixels, the weight etc. We will refer to both ratio and interval domains as numerical domains. In real data, most frequent are nominal and interval/ratio domains. Ordinal domains are created with a user influence -a user will say that every object that weights more than 600g is heavy, above 200g is medium and less than 200g is light.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Boolean preference</head><p>Boolean user preference model is used in some methods, where the user preference model is not explicitly mentioned. This model distinguish only two states of an object -either preferred or non-preferred. This is very simple approach with little semantic, but may be used when lot of computing power is required. In these cases, preference is represented by a vector of n bits, we will note this vector v(u) : U → [0, 1] n . Operations on these vectors are fast when only binary operations like and, or, xor etc. are needed. These operations may be sufficient for some inductive methods but not for others. For example, this model can be successfully used in the collaborative filtering, which is described in Section 3.2.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Comparison of two user preferences.</head><p>Computing the similarity of two users may be computed in following way</p><formula xml:id="formula_0">s(u 1 , u 2 ) = 1 − n i=1 (v(u 1 ) xor v(u 2 ))[i]/n.</formula><p>The sum in equation expresses the number of ratings, where both user disagree, e.g. object is preferred for u 1 and non-preferred for u 2 . If we based the similarity on common preferences of u 1 and u 2 , it will be influenced by the number of rated objects, which is not desirable. If this fact is of no relevance, alternative way for computing similarity may be</p><formula xml:id="formula_1">s(u 1 , u 2 ) = n i=1 (v(u 1 ) and v(u 2 ))[i]/n.</formula><p>Computing the similarity of two users is essential for methods like collaborative filtering. Surprisingly, there are not many articles concerning this problematic.</p><p>Further investigation and research in this area may reveal interesting ideas.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Preference relations</head><p>Preference relations are the oldest model of user preferences models, its description may be found e.g. in <ref type="bibr" target="#b14">[15]</ref>. We left out the case when o 1 is a little better than o 2 . We may create new relation</p><formula xml:id="formula_2">Q, so that Q(o 1 , o 2 ) states that o 1 is a little better than o 2 . By a simple extension, set of relation Q 1 , ..., Q n will represent the fact that o 1 is a little better than o 2 , with Q 1 representing the lowest difference of preference of o 1 and o 2 and Q n the highest difference.</formula><p>Properties of relations determine properties of preferences. There are several properties, such as the existence of a minimum or the completeness (linearity) of the relation. For deeper insight in these properties, we again recommend <ref type="bibr" target="#b14">[15]</ref>, which is specialized survey of preference relations.</p><p>All these structures may be extended to valued structures. One special case is many valued logic, studied more in depth in 2.3. An example of a valued structure is µ(P (a, b)) : O 2 → [0, 1]. The interval [0, 1] may be replaced by any other linear numerical structure, and it represents the degree (truth value) of the relation. Valued relations may successfully replace relations Q 1 , ..., Q n , which are a middle step between standard relations and valued relations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Comparison of two user preferences.</head><p>When we want to compare preferences of two users, we have to compare two preference relations. When relation is ordered linearly, we compare two ordered lists of objects. In that case, we may count the number of permutation between both lists, which is traditional measure of computing the similarity (or the distance, in this case). However, this may be not the best distance used, because it makes no difference of distance between permuted objects. Switching two neighbour objects makes less change than switching first and last object.</p><p>In <ref type="bibr" target="#b13">[14]</ref> the distance of two interval fuzzy preference relations is described. However, it can't be simply used as a generalization of simple preference relation, because it will degrade into simple 'equal' or 'not equal' semantics.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3">Many valued logic</head><p>Many valued logic is an extension of the traditional two valued logic. In the two valued logic, a variable may be either true or false, in the many valued logic a whole set of possible truth values is introduced, often denoted as T . T should form a lattice, typical case is a linear structure and the most used is interval [0, 1]. The set T will represent the set of preference values where 1 is most preferred and 0 is least preferred. Other structures than [0, 1] are possible to use, for example discrete set {Worst, Worse, Neutral, Better, Best} or {One star, Two stars, Three stars, Four stars, Five stars} may be relevant in some cases. We will interpret the truth value as a degree of preference.</p><p>When creating this extension of the two valued logic, we must define a new semantics for logical operators, predicates and quantifiers. This definition can be found for example in <ref type="bibr" target="#b12">[13]</ref>, but this is not of major interest in this work.</p><p>We will use two structures from the many valued logic -the first are fuzzy functions of an attribute domain and the second is an aggregation function. A fuzzy function represents user's ordering and normalization of a domain. For example, consider the attribute price. Most people prefer low prices over high prices. In figure <ref type="figure" target="#fig_1">2,a</ref>) we may see an example of the fuzzy function for price. An aggregation function is used to aggregate several truth values, or preferences, into one truth value, therefore it is a function @ : T m → T . There are few restrictions on an aggregation function -it must be monotone in all variables, and @(0, .., 0) = 0 and @(1, .., 1) = 1. An aggregation function is very suitable for the modeling of the user preferences of the complex objects. The user aggregates attributes of an object into the preference of the object itself. An example of an aggregation function may be</p><formula xml:id="formula_3">@ U1 (MPix U 1 (x), Fast U 1 (x), Cheap U 1 (x)) = 5 * MPix U 1 (x) + 1 * Fast U 1 (x) + 3 * Cheap U 1 (x)<label>9</label></formula><p>Symbols MPix, Fast and Cheap denotes fuzzy sets of particular attributes. E.g. Cheap U 1 (D50) represents how camera D50 is cheap for U 1 .</p><p>We consider the weighted average as a good example of a user fuzzy function. It has clear semantics, because we can see directly, which of the attributes are important for the user and which are not. Even more, from the weight we can deduce how much important an attribute is. However, many more aggregation functions that fits better to psychological aspects of human decision process may be represented. An example of a more complicated aggregation function is in Figure <ref type="figure" target="#fig_1">2,b</ref>). These two mechanisms allow us to create very flexible model of user preferences and moreover, the aggregation function models also user decision process.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Inductive methods</head><p>In this section, we will examine several inductive methods that are used to create a user preference model from some input. The user preference model is often independent of the inductive method, the input may be represented in several ways but often the paradigm of object with some attributes is expected.</p><p>Most of the methods expect a training set of objects, which are supposed to belong to 'classes'. These classes of objects in the training set may have different forms, depending on the model of user preferences we are using. For example, when using many valued logic, one class may be o : P (o) ∈ [0, 0.1]. Some of the models of user preferences we have described above do not have a direct interpretation as classes. With preference relations, we have only comparison between two objects. We assume that a method will transform user preferences into several (possibly discrete) classes. For preference relations we may order objects and associate a weight corresponding to the position in the ordered set.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Inductive logic programming</head><p>Inductive logic programming is a method to obtain a logic program. This program may be very general, in our case, it will represent rules that user uses in decision process. After application of these rules to an object, the preference of that object should be obtained. We will describe only predicate logic programs, which are more expressive than sentential logic programs. An introduction to induction of logic programs may be found in <ref type="bibr" target="#b10">[11]</ref>.</p><p>Rules have a head and a body. The head of a rule is a single predicate and the body is a conjunction of predicates. When using fuzzy predicate logic, each predicate has also value that represents the truth value of the predicate. For simplicity (and because of space limitation), we will describe two-valued logic program.</p><p>For example, following rules may represent user preferences of cameras GoodMPix(camera) &lt;-MPix(camera)&gt;5; GoodWeight(camera) &lt;-Weight(camera)&lt;700 &amp; Weight(camera)&gt;300; GoodCamera(camera) &lt;-GoodMPix(camera) &amp; GoodWeight(camera);</p><p>Inductive process works with this input QH is a set of candidates to hypothesis and R is a set of rules, which transform H. An example of a rule may be dropping a clause or adding a clause to the body of H. During each step, a hypothesis H and rules that will be applied to H are chosen. Result of the application of rules on H are then stored in QH and candidate set is pruned. Pruning means that useless candidates are deleted. Implementation of each of methods Delete, Choose, Prune and Stopcriterion may be different. Also the set of available rules R may differ across implementations.</p><p>An example of application of a rule on a hypothesis GoodCamera(D2x) &lt;-; may be GoodCamera(camera) &lt;-Megapixels(camera)=12; or GoodCamera(camera) &lt;-Weight(camera)=1150; This is an example of a generalization rule, whose result must be verified on E + and E − .</p><p>The hypothesis should be completely correct, i.e. it should have the Posterior sufficiency property. However, if we relax this property, a kind of probabilistic rules will be generated. The probabilistic approach is further studied in 3.4.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Collaborative filtering</head><p>Collaborative filtering represent widely used method for acquisition of user preferences. It is based on assumption, that the preference of user u 0 on object o will be the same as the preference of users u 1 , ..., u k that are 'similar' to u 0 . The similarity of users is based on similarity of ratings on objects. Many collaborative filtering methods are described in <ref type="bibr" target="#b9">[10]</ref>.</p><p>This method requires a lot of ratings of objects by a lot of users. For computing the similarity of users, we need a lot of object ratings, for accuracy of computing the rating of object o, we need a lot of users similar to u 0 .</p><p>There are several different algorithms for collaborative filtering. The first and most simple is K-NN, the K nearest neighbor. This is most intuitive algorithm -for a user u 0 , we find the K nearest users u 1 , ..., u K . The distance is computed by the similarity of users' ratings, for example</p><formula xml:id="formula_4">s(u i , u j ) = n l=1 (P i (o l ) − P j (o l )) 2<label>(1)</label></formula><p>Having these K nearest neighbours, we may compute the rating of objects as average of users' ratings</p><formula xml:id="formula_5">P 0 (o i ) = K j=1 (P j (o i ))/K<label>(2)</label></formula><p>Another method of computing new ratings is to use a naive Bayes classifier <ref type="bibr" target="#b5">[6]</ref>.</p><p>For each object o, we construct a separate Bayes classifier. Input of the network are the ratings of all objects other than o. Bayes network will answer the question "What is the value P (o), when the user rated other objects this way?". Bayes network learns its parameters from the ratings of users that have rated o.</p><p>Other methods are considered as a content filtering methods -they work with the objects and their properties rather then with the preferences of other users. However, both approaches are often combined. Collaborative filtering can not be appropriately used for new objects, which have not yet been rated by any user. For this reason, some kind of the content filtering is also used and the results of both methods are combined together.</p><p>However, some of the presented methods may be used both on other users' ratings and the properties of objects, or both together.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Decision trees and rules</head><p>Decision trees are well known structure from the data mining theory, they are used to model user decision process (or any decision process). Decision trees are oriented trees with class names in leaves and a rule associated to each inner node. An example of a decision tree which models user preferences of digital cameras is in figure <ref type="figure">3</ref>. We can see that cameras with at least 10 megapixels are good, cameras that costs more than 500$ are bad.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Fig. 3. An example of a decision tree</head><p>Theory of the induction of the decision trees may be found in <ref type="bibr" target="#b11">[12]</ref>. Induction of a decision tree is discussed mainly for discrete attributes with few values. Decision trees were extended to fuzzy decision trees, one of a new contribution is in <ref type="bibr" target="#b4">[5]</ref>, older one is <ref type="bibr" target="#b2">[3]</ref>.</p><p>Basic induction procedure starts with a 'training set', which is a set of objects with known classes. From these objects, a tree is constructed and then it is verified on a test set of objects, which is also a set of objects with known attributes. The construction is typically a top-down algorithm. During each step, all possible splits are considered and the most appropriate one is chosen. The appropriateness is measured with an 'impurity measure'. Impurity measure measures how evenly the data are spread in classes. When all objects are only in one class, impurity measure is 0, when all classes contain same number of objects, impurity measure should be 1. Most used is entropy-based impurity measure ( <ref type="bibr" target="#b11">[12]</ref>) and Gini index ( <ref type="bibr" target="#b3">[4]</ref>). The structure of a tree depends mainly on the impurity measure used during its construction.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4">Probabilistic methods</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Probabilistic methods use probability as a method of inducing user preference.</head><p>There are several possible approaches to statistical interpretation of preference data, e.g. <ref type="bibr" target="#b8">[9]</ref>. Usage of probability is reasonable when working with user preferences, because few users have consistent preferences. Often, an exception from a general rule occurs. This exception have to be handled explicitly in nonprobabilistic methods, but it creates no problem in the probabilistic methods. For example in inductive logic programming, we can assign a probability to each rule, denoting how much the rule is true in general, or in decision trees the probability would be associated to each left node.</p><p>Probabilistic model for boolean preference model. The probabilistic preference model proposed in <ref type="bibr" target="#b8">[9]</ref> is based on two measures -the first is the actual user preference and the second is the accessibility (or the frequency) of the object. The second one is important because when trying to induce user preferences from user behaviour it is apparent that user will rather examine frequent items than rare items just because they are more frequent.</p><p>The preference of an object is P ref</p><formula xml:id="formula_6">(x) = f (x|V )</formula><p>, where V is a user profile or a user history and f returns the preference value of object x. Objects have again several attributes, in <ref type="bibr" target="#b8">[9]</ref> called 'features'. The preference of an object is computed as</p><formula xml:id="formula_7">P ref (x) = 1/|X| a∈X P ref (a),<label>(3)</label></formula><p>e.g. it uses the average of attributes values preferences. The problem of finding P ref (a) is then analyzed, actually in the similar way as in our previous work <ref type="bibr" target="#b6">[7]</ref>. We will compare these approaches in Section 3.4. The suggestion in <ref type="bibr" target="#b8">[9]</ref> is to use formula</p><formula xml:id="formula_8">P ref (a) = I(X(a); V ) = log P (X(a)|V ) P (X(a))<label>(4)</label></formula><p>where X(a) is the set of objects containing a. In other words, attribute value a is preferred, if the probability that the user selects an object with a is higher than the probability of the occurrence of an object with a in the whole set of objects.</p><p>Probabilistic model for many valued logic. A probabilistic model for the many valued logic is our contribution. It is aimed on nominal attributes and uses only weighted average as aggregation function. We concentrate on the case, when we know preferences of some objects and user aggregation function but not the preference of attribute values. We are missing the preference of an attribute value a which is the value of attribute A k . But we know the preferences of objects and the aggregation function. We will consider the set X(a) of objects which have the attribute value a. We will look into the distribution of the preference of these objects. When most of the objects in X(a) have high preference, attribute value a will also have high preference. Formally,</p><formula xml:id="formula_9">P (a) = o∈X(a) P (o) | {o ∈ X(a)} | . (<label>5</label></formula><formula xml:id="formula_10">)</formula><p>The ratio between the weight of the attribute A k in aggregation function @ and the sum of the weights of all attributes represents the probability that the preference is computed correctly. It is denoted formally in the following equation</p><formula xml:id="formula_11">P (A k ) = W (@, A k ) i=1,..,m W (@, A i ) , (<label>6</label></formula><formula xml:id="formula_12">)</formula><p>where W (@, A j ) is the weight of attribute A j in @. The preference of an object is influenced more by an attribute with a big weight than by an attribute with a small weight. Therefore this method is useful mainly for the attributes with big weight.</p><p>Computing preference of one attribute value a may be costly when the number of objects with a is big. However, higher number of objects with a means also higher precision of this method. Naturally, this method is only useful for the domains with discrete values, especially non ordered domains like color or manufacturer. This method can't be successfully used for continuous domains, because there will be very few objects with the same attribute value. However, we may divide these continuous domains to a set of discrete intervals, and use the method proposed above on these intervals.</p><p>Comparison with our model. Our model is an extension over the model proposed in <ref type="bibr" target="#b8">[9]</ref>. There are two aspects in which our approaches differ 1. In <ref type="bibr" target="#b8">[9]</ref> the boolean user model is used (implicitly). We use many-valued logic model, which is more general. 2. The preference of an object is computed in <ref type="bibr" target="#b8">[9]</ref> as a simple average of preference of its attributes. In our model, weighted average is used.</p><p>However, there is a similarity in our approaches -we use the preference of objects for acquiring the preference of attribute values. This is an inverse process to deduction, where the preference of an object is computed from the preference of its attribute values.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Conclusion</head><p>In this paper, we have reviewed some of the main user preference models. There are other models as well, their complete listing is beyond the scope of this paper, we recommend <ref type="bibr" target="#b14">[15]</ref> to the interested reader. The user model is used in a web system to better present data to the user or to alter his/her query in order to the results of the query actually better fit his/her preferences.</p><p>The creation of the user model is often done by inductive methods, which were studied in this paper in Section 3. We presented methods that are used for induction of user preferences and one probabilistic model for boolean user preferences. We have developed a similar approach for many valued logic, which is more general and flexible than the method studied in section 3.4. The precision and the computational cost of our approach is still to be tested on real data. These experiments are however beyond the scope of this paper, which is an overview of methods used for the induction of the user preferences.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Complex system for user preferences on the web</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. Fuzzy function of price (a) and an example of a more complicated aggregation function (b)</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head></head><label></label><figDesc>Basic idea behind preference relations is to characterize the relation between objects o 1 and o 2 . We can say that o 1 is more preferred than o 2 , o 1 is equal to o 2 , o 1 is incomparable to o 2 or that o 1 is a little better than o 2 but not much. For the strict preference, we traditionally denote this relation as P . Then P (o 1 , o 2 ) states that o 1 is more preferred than o</figDesc><table /><note>2 . For equivalence of two objects, we use I, e.g. I(o 1 , o 2 ) means that o 1 is as preferred as o 2 . Finally, relation R is created as union of P and I, R(o 1 , o 2 ) meaning o 1 is equal or preferred to o 2 . For incomparability, relation J is introduced. Then J(o 1 , o 2 ) means that o 1 and o 2 are incomparable.</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head></head><label></label><figDesc>1. The background theory B. 2. Positive examples E + . 3. Negative examples E − . Background theory is used to infer new statements H (hypothesis) about examples. Both E + and E − are formulas, but E − have empty head, e.g.</figDesc><table><row><cell>&lt;-GoodCamera(D50);</cell><cell></cell></row><row><cell>&lt;-GoodCamera(D40);</cell><cell></cell></row><row><cell cols="3">On the other side, positive examples have empty body, e.g.</cell></row><row><cell>GoodCamera(D200) &lt;-;</cell><cell></cell></row><row><cell>GoodCamera(D2x) &lt;-;</cell><cell></cell></row><row><cell cols="3">We present also an example of a background knowledge B:</cell></row><row><cell>Weight(D2x)=12 &lt;-;</cell><cell></cell></row><row><cell cols="2">Megapixels(D2x)=1150 &lt;-;</cell></row><row><cell cols="2">Four conditions must be fulfilled</cell></row><row><cell cols="2">1. Prior satisfiability B &amp; E −</cell></row><row><cell cols="3">2. Posterior satisfiability B &amp; E − &amp; H</cell><cell>.</cell></row><row><cell>3. Prior necessity B</cell><cell>E + .</cell></row><row><cell cols="2">4. Posterior sufficiency B &amp; H</cell><cell>E + .</cell></row><row><cell cols="3">Now we will describe a general algorithm of hypothesis construction, as pro-</cell></row><row><cell>posed in [11].</cell><cell></cell></row><row><cell>QH = Initialize();</cell><cell></cell></row><row><cell>do</cell><cell></cell></row><row><cell>Delete H from QH;</cell><cell></cell></row><row><cell cols="3">Choose rules r 1 , ..., r k ∈ R to be applied to H;</cell></row><row><cell cols="3">Apply r 1 , ..., r k to H, obtaining H 1 , ..., H n ;</cell></row><row><cell cols="2">Add H 1 , ..., H n to QH;</cell></row><row><cell>Prune QH;</cell><cell></cell></row><row><cell cols="2">while not Stop-criterion(QH)</cell></row></table><note>The symbol represent the contradiction or false. The meaning of these condition is clear -with B and E − we should not get a contradiction, e.g. E − should comply to the background knowledge. With the B, E − and H we should not get a contradiction either. On the other hand, we want that examples are not deducible from B itself, only with addition of H.</note></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgment</head><p>Supported by Czech projects MSM 0021620838, 1ET 100300517 and 1ET 100300419.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<ptr target="http://www.w3.org/TR/owl-features/" />
		<title level="m">Owl, ontology web language</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<ptr target="http://www.w3.org/TR/rdf-primer/" />
		<title level="m">Rdf data format</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Learning fuzzy decision trees</title>
		<author>
			<persName><forename type="first">B</forename><surname>Apolloni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Zamponi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">M</forename><surname>Zanaboni</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Neural Networks</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<biblScope unit="issue">5</biblScope>
			<biblScope unit="page" from="885" to="895" />
			<date type="published" when="1998">1998</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m" type="main">Classification and Regression Trees</title>
		<author>
			<persName><forename type="first">L</forename><surname>Breiman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Friedman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Olshen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Stone</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1993">1993</date>
			<publisher>Chapman &amp; Hall</publisher>
			<pubPlace>New York</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">Supervised Ranking, from semantics to algorithms</title>
		<author>
			<persName><forename type="first">K</forename><surname>Cao-Van</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2003">2003</date>
		</imprint>
		<respStmt>
			<orgName>Ghent University</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Ph.D. dissertation</note>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">On the optimality of the simple bayesian classifier under zero-one loss</title>
		<author>
			<persName><forename type="first">P</forename><surname>Domingos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">J</forename><surname>Pazzani</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Machine Learning</title>
		<imprint>
			<biblScope unit="volume">29</biblScope>
			<biblScope unit="issue">2-3</biblScope>
			<biblScope unit="page" from="103" to="130" />
			<date type="published" when="1997">1997</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Methods for finding best answer with different user preferences</title>
		<author>
			<persName><forename type="first">A</forename><surname>Eckhardt</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Czech only</title>
				<imprint>
			<date type="published" when="2006">2006</date>
		</imprint>
		<respStmt>
			<orgName>Charles University in Prague</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Master&apos;s thesis</note>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Jonas Lundberg Design Perspectives</title>
		<author>
			<persName><forename type="first">Lars</forename><surname>Hult</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Magnus</forename><surname>Irestig</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Human-Computer Interaction</title>
		<imprint>
			<biblScope unit="volume">21</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="5" to="48" />
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">A statistical model for user preference</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">Y</forename><surname>Jung</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J.-H</forename><surname>Hong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T.-S</forename><surname>Kim</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on</title>
		<imprint>
			<biblScope unit="volume">17</biblScope>
			<biblScope unit="issue">6</biblScope>
			<biblScope unit="page" from="834" to="843" />
			<date type="published" when="2005-06">June 2005</date>
		</imprint>
	</monogr>
	<note>Knowledge</note>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">Collaborative filtering: A machine learning perspective</title>
		<author>
			<persName><forename type="first">B</forename><surname>Marlin</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2004">2004</date>
		</imprint>
		<respStmt>
			<orgName>University of Toronto</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Master&apos;s thesis</note>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Inductive logic programming: Theory and methods</title>
		<author>
			<persName><forename type="first">S</forename><surname>Muggleton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">D</forename><surname>Raedt</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Logic Programming</title>
		<imprint>
			<biblScope unit="volume">19</biblScope>
			<biblScope unit="issue">20</biblScope>
			<biblScope unit="page" from="629" to="679" />
			<date type="published" when="1994">1994</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Induction of decision trees</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">R</forename><surname>Quinlan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Mach. Learn</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="81" to="106" />
			<date type="published" when="1986">1986</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Fuzzy logic programming</title>
		<author>
			<persName><forename type="first">P</forename><surname>Vojtáš</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Fuzzy Sets and Systems</title>
		<imprint>
			<biblScope unit="volume">124</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="361" to="370" />
			<date type="published" when="2001">2001</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">On compatibility of interval fuzzy preference relations</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Xu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Fuzzy Optimization and Decision Making</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="217" to="225" />
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<title level="m" type="main">Preference modelling. Multiple Criteria Decision Analysis: State of the Art Surveys</title>
		<author>
			<persName><forename type="first">M</forename><surname>Öztürké</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Tsoukias</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Vincke</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2006">2006</date>
			<publisher>Springer</publisher>
			<pubPlace>New York</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
