<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Improving Novelty in Streaming Recommendation Using a Context Model *</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Alexandra</forename><surname>Doina</surname></persName>
						</author>
						<author role="corresp">
							<persName><forename type="first">Simone</forename><surname>Dumitrescu</surname></persName>
							<email>doina.dumitrescu@uam.es</email>
						</author>
						<author>
							<persName><surname>Santini</surname></persName>
						</author>
						<author>
							<affiliation key="aff0">
								<orgName type="institution">Escuela Politécnica Superior</orgName>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff1">
								<orgName type="institution">Universidad Autónoma de Madrid Madrid</orgName>
								<address>
									<country key="ES">Spain</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Improving Novelty in Streaming Recommendation Using a Context Model *</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">88A6680881D1DAD11D40B54A2DDD31F3</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-23T23:07+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>H.3.3 [Information Systems]: Information Search and Retrieval Algorithms</term>
					<term>Metrics</term>
					<term>Experimentation Novelty</term>
					<term>Diversity</term>
					<term>User Model</term>
					<term>Context</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In recent years there has been an increasing research interest in novelty/diversity detection in Information Retrieval and in Recommendation Systems. We propose a model that increases the novelty of recommendations using a context user profile that was created automatically using self-organizing maps. Our system was evaluated on the Reuters Corpus Volume 1 and our experiments show that filtering the recommended items using a novelty score derived from the contextbased user profile provides better search results in terms of novel information that is presented to the user.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">INTRODUCTION</head><p>The plethora of information available on the internet about virtually anything, and the duplication of this information due to the significant replication of the sources has given a great relevance to the issues of novelty and diversity in retrieval and recommendation systems <ref type="bibr" target="#b2">[2,</ref><ref type="bibr" target="#b5">5]</ref>. The definition of novelty and diversity generally accepted in information retrieval is that diversity helps the server deal with query ambiguity, while novelty helps the user deal with query underspecification <ref type="bibr" target="#b8">[8]</ref>. Consider a query composed of the word manhattan. The query is ambiguous because the server doesn't know whether the user wants to know something about the neighborood in New York, about the cocktail, or about the Indian tribe: the query has several possible interpretation, and whoever formulated the query is interested in general in only one of them. Without further specification, the safest bet for the server is to provide a diverse result set, that is, one that covers all these topics, possibly in an amount proportional to the estimated a priori user interest (one can assume, for instance, that more people are interested in New York than in the Indian tribe). Even if the query is not ambiguous and the results are about what the user really meant (say, the neighborhood in new york), the user would want every document to be novel, that is, to provide information that is not present in other documents. A highly informative text about the history and the human landscape of Manhattan is probably a very good first result, but another document almost identical to the first is not an equally desirable second result: the information that it contains has already been seen. That is, every interpretation of a query contains several aspects, and the person who formulated it will be interested, to some degree, in all of them. These considerations highlight an important difference between diversity and novelty. Diversity is something that the user wouldn't really want: the person interested in the manhattan cocktail would be very happy to receive results only about the cocktail. Diversity is needed mainly by the server, which doesn't know enough about the user true intentions. Novelty, on the other hand, is something the user wants, since she will be interested in all the aspects of her query. Novelty is a way to avoid redundancy, while diversity is a way to deal with the server's ignorance.</p><p>If we have access to a model of the user's interests, then it is possible to reduce ambiguity and, therefore, to reduce the need for diversity, concentrating only on increasing novelty. We present a method for increasing novelty that relies on a context model built on texts previously seen by the user. The scenario that we use here to exemplify the working of our method is that of the recommendation of news, and the context is composed of past news that the user has considered interesting. This is just one possible way in which context can be gathered. Our model is very general and can be used to represent text or multimedia data. In past work, we have used, for example, the contents of the user's hard disk to represent his main interests <ref type="bibr">[7]</ref>. In this paper we show and evaluate a modification to our model with the purpose of increasing the context coverage of the recommendations, that is, the fraction of the user context that the recommended items will cover, trying to avoid items that refer to parts of the context already covered by previously seen items.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">STREAM RECOMMENDATION</head><p>One important distinction between this work and others on novelty and diversity in recommendation is that we are dealing with a stream of news, that is, with a continuouspotentially infinite-interaction, while typical recommendation systems work based on sessions: the user chooses an item or requests a recommendation, and a finite set of results is shown. This operation may be repeated several times, as the user selects several items and receives correspondingly changing recommendations. No work, to the best of our knowledge, has considered novelty in a session-less situation, in which a countable stream of items arrive and a system has to decide whether they are relevant for the user at that moment. Note that, differently from the case of a closed session, we don't have access to the whole data base of items that can be recommended. We don't even have access to the items that have already been shown, since in a stream their number is unbound. All we can do, at any given time, is take a decision only considering how the past history of items has modified the user context. The system is shown schematically in figure <ref type="figure">1</ref>. We have a stream of items (news, in our test scenario) that arrive continuously to the user. The user is described by the context model C (the details of this representation will be given in the next section) and each news is compared with the context and assigned a score. Whenever the score is above a certain threshold (or a predefined criterion is met) the item is presented to the user. At the same time, the context undergoes a first kind of change, which we call a short term change. This change (used to improve the novelty of the recommended items) is characterized by a time constant β. After a time, depending on β, the change provoked by any individual item disappears. This is one important difference between our model and the session based ones. Let us assume that a person is interested in politics. News about the election of the new French president will undoubtedly be of interest to her. However, she will probably not be interested in receiving many similar news about the election one after the other. The effect of the short term change will be do "deactivate", for a time, the part of the context that deals with this kind of news, so that further news about the French elections will not, coeteris paribus, be considered as interesting. However, this situation should not be permanent: after a while (a few hours, a day,...), further news about the French President will again be considered interesting, and should be recommended.</p><p>In addition to this short term change, the context evolves in a permanent and irreversible way, again under the influence of the documents read (as well as, possibly, based on other elements such as the documents that the user creates on her computer, or the emails that she writes). A number of strategies can be used for this evolution: the user can be asked to mark certain documents of special interest, the system can present only a summary, and all the documents whose complete text is accessed are permanently entered into the context, etc. We have considered the evolution of the context based on documents of interest elsewhere <ref type="bibr">[7]</ref>, and we shall not consider it here. From the point of view of this paper, therefore, the context, defined as in the following section, is fixed, and only the short term changes, needed to increase novelty, are considered.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">THE USER MODEL</head><p>There is convincing evidence that a user model can help attain a higher precision in information retrieval and recommendation systems <ref type="bibr" target="#b1">[1,</ref><ref type="bibr" target="#b6">6]</ref>. In this paper, we are interested in using such a model to achieve novelty, that is, to obtain recommendations that span the whole gamut of the user's interests. The context model that we use was presented (without the extension to enforce novelty that we present here0 in <ref type="bibr">[7]</ref>, to which the reader is referred for details; here we shall only give a brief description. The basis of the model is a set of documents, which can be composed of the documents on which the user is working at the time or, in our case, by news items marked in the past. The model doesn't change depending on the source of the documents. In the following, in order to make the tests self-contained, we shall always assume that the context is based on a set of news items that supposedly the user has seen in the past and that he has found relevant.</p><p>All the documents in the contexts are considered as a single "macro-document" and processed consequently. On this macro-document we perform stop-word removal and stemming. Each term t is assigned a weight wt using tf-idf weighting <ref type="bibr" target="#b1">[1]</ref>. We use the standard vector model representation of information retrieval in which each different term corresponds to an axis in the word space. We consider n-grams, constituted of n consecutive terms that appear in the document. If the n-gram p k is composed of the terms tu 1 , . . . , tu n , with weights wu 1 , . . . , wu n , then its representation in the word space is given by the point In the word space, we lay a self-organizing map, using a modification of WEBSOM <ref type="bibr" target="#b3">[3]</ref>. The map is a grid of elements called neurons, each one of which is a point in the word space and is identified by two integer indices, that is, a neuron is given as:</p><formula xml:id="formula_0">p k = 1 n i=1 w 2 u i (. . . ,</formula><formula xml:id="formula_1">[µν] = ([µν]1, . . . , [µν]T ) 1 ≤ µ ≤ N, 1 ≤ ν ≤ M<label>(1)</label></formula><p>The map is discrete, two-dimensional<ref type="foot" target="#foot_0">1</ref> with the 4-neighborhood topology. The neurons are immersed in the T -dimensional word space ([µν] ∈ R T ), and in our version of WEBSOM their weights are normalized. That is, throughout the algorithm we enforce T t=1 [[]µν]<ref type="foot" target="#foot_1">2</ref> t = 1 for all map indices µν. The neurons are at the same time elements of the discrete grid and points in S T −1 . As points in the discrete grid, the relevant distance between two neurons is their graph distance:</p><formula xml:id="formula_2">δ([ζξ], [µν]) = |ζ − µ| + |ξ − ν|<label>(2)</label></formula><p>As points in S T −1 , it is possible to determine the similarity between a neuron and any other point p ∈ S T −1 as</p><formula xml:id="formula_3">s([µν], w) = T i=1 [µν]ipi<label>(3)</label></formula><p>On this map we define a neighborhood function, h(t, n), which depends on two parameters t, n ∈ N; n is the graph distance between two given neurons, t is a time parameter that increases as learning proceeds. The function h(t, n) represents the "degree of neighborhood-ness" of two neurons at a distance n at time t; We assume that 0 ≤ h(t, n) ≤ 1, h(t, 0) = 1, and that h is monotonically decreasing in n and t. The degree to which neuron [ζξ] belongs to the neighborhood of neuron [µν] at time t is given by h(t, δ([ζξ], [µν])).</p><p>In addition to the neighborhood we define a positive learning parameter α(t), t ∈ N, monotonically decreasing with t.</p><p>In order to create the map for a context C, all the points in it are presented to the map, and the training algorithm is applied. We call the presentation of a point p ∈ C an event of learning, and the presentation of all the points of C an epoch. Learning consists of a number of epochs, counted by a counter t. The neurons of the map are at first spread randomly in the word space; then, for each event consisting of the presentation of the point p, the neuron with the highest similarity to p is found:</p><p>[ * ] = arg max </p><p>the neuron [ * ] and all its neighbors are shifted towards p. The amount of this shift depends on the learning parameter α and on the distance from [ * ] on the map:</p><formula xml:id="formula_5">∀[µν] [µν] ← [µν] + α(t)h(t, δ([ * ], [µν])) • (p − [µν])<label>(5)</label></formula><p>Finally, all neurons are re-normalized in order to maintain them on the sphere S T −1 :</p><formula xml:id="formula_6">∀[µν] [µν] ← [µν] i [µν] 2 i (6)</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">NOVELTY IN NEWS FILTERING</head><p>If we don't consider novelty, we can filter the news items as they arrive simply by representing them as a point u ∈ S T −1 (considering it as a document and using the same information retrieval techniques that we have used to represent the context), and determining the maximal similarity between u and the neurons of the map</p><formula xml:id="formula_7">r = max [µν] s([µν], u)<label>(7)</label></formula><p>the items that yield a value of r below a certain threshold are considered interesting and presented to the user. As we have mentioned in the introduction, the problem with this simple solution is that very similar news items can have high score because they are similar to the same neuron (that is: to the same part of the user context), although the presentation of more than one adds little to the information content of the first one. To increase the diversity of the results, we rewrite <ref type="bibr">(7)</ref> adding an interest factor wµν to each neuron, and computing the relevance of the news item u as</p><formula xml:id="formula_8">r = max [µν] wµν s([µν], u)<label>(8)</label></formula><p>Let us call [ * ] the "winning" neuron for a given item (the neuron closest to it, viz. the neuron for which the maximum of eq. ( <ref type="formula" target="#formula_8">8</ref>) is attained). Whenever an item is recommended, the interest factor of each neuron is updated according to the equation</p><formula xml:id="formula_9">wµν ← λwµν if [µν] = [ * ] min{1, (1 + β)wµν } otherwise (9)</formula><p>with 0 &lt; λ &lt; 1 and β &gt; 0. According to this equation, the interest of the neuron closest to a recommended news item is reduced by a factor λ, reducing its contribution to ( <ref type="formula" target="#formula_8">8</ref>) and therefore making it less likely to win again. This means that further news items close to the one that caused this neuron to win will be less likely to be recommended. If the neuron is no longer the winner for any item, its interest will be increased by a factor (1 + β) with each recommended item, until the value 1 is restored (or until the neuron wins again, in which case its interest factor will again be reduced). If the neuron wins only once, its relaxation time (the number of items necessary for it to go back to 1) is</p><formula xml:id="formula_10">tr = − log λ log(1 + β)<label>(10)</label></formula><p>If we want to achieve effective diversity, this time should be of the order of magnitude of the number of neurons in the network, but not greater, as if neurons win too often their interest factor will decrease, on average, as a function of time. A reasonable choice is to set tr ≈ K 10 , where K is the number of neurons. Expressing, say β as a function of λ and tr, we obtain</p><formula xml:id="formula_11">β = 1 λ 1 tr − 1 (11)</formula><p>so, for the desired recovery span, we have</p><formula xml:id="formula_12">β = 1 λ 10 K − 1 (12)</formula><p>In this schema, we only reduce the interest factor of the neuron [ * ]. In our tests, we also tried another schema, in which we reduce (by a progressively smaller amount) the interest factor of neurons in the neighborhood of the winner using the same function h that we have used in the training algorithm. In this case, the update of the interest factor is done with the function</p><formula xml:id="formula_13">wµν ← λh(0, [µν], [ * ])wµν if δ([µν], [ * ]) &lt; τ min{1, (1 + β)wµν } otherwise (<label>13</label></formula><formula xml:id="formula_14">)</formula><p>where h is the neighborhood function, and τ is a suitable threshold (in our experiments, we set τ = 2).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">TESTS</head><p>We consider two measures. The first one is the standard precision (the fraction of recommended documents that are relevant). Given a context and a set of recommendations, we expect that the use of novelty will somewhat decrease precision (if there are two very relevant documents that are almost identical, a system that considers novelty will recommend only one of them). So, this measure is a sort of quality control: we do accept a certain reduction of precision, but we want to keep it under control, to avoid losing too much result quality. The second measure is a determination of novelty. We measure, for a given number of recommended items, the number of neurons that win. If D is the number of documents recommended and, during the recommendation of these documents, nw different neurons are winners (i.e. they are the neurons that achieve the maximum in eq. ( <ref type="formula" target="#formula_8">8</ref>)), then the coverage is defined as V = nw/D (0 &lt; V ≤ 1). Note that nw &lt; K so that if the documents arrive in a stream (D → ∞) in the long run this value will tend to zero. In order to obtain a significative value, we always consider sets of recommended documents smaller than the number of neurons in the network.  To validate the proposed approach, we conducted some experiments using 2.5 gigabytes of uncompressed news stories from Reuters Volume 1 (RCV1-v1) <ref type="bibr" target="#b4">[4]</ref>. The Reuters's is a collection of 806,791 news stories, each one associated with some metadata. In our experiments, we used the text and the topics categories each news document belongs to. The RCV1-v1 test collection contains 117 topics in a three-level hierarchy with, at the top, four categories: CCAT (Corporate/Industrial), ECAT (Economics), GCAT (Government/Social), MCAT (Market). Fourteen of these categories contain no document, and in our experiments we used only the 103 that do.  We began by creating a context. To this end, we used 5% of the news items contained in the sub-categories (of MCAT) M11 (EQUITY MARKETS) and M13 (MONEY MARKETS), for a total of about 5000 documents. Once the context is in place, we create a stream of news with all the elements in the Reuters data base and use our method for recommending items that are close to the context. As a ground truth, we consider as relevant all the items that are in the context categories (M11 and M13) and in their parent category (MCAT), but not those that are in the sibling categories of M11 and M13. We stop the stream when 500 items have ben recommended (this is necessary in order to get significant coverage values, as we mentioned before). Each item is analyzed and represented as a point u ∈ S T −1 (see section 4). The weights of the terms that compose each item are determined according to two different schemas: Simple weighting (without an inverse document frequency (idf) term; the weight of a word is calculated using the normalized term frequency (tf ), and tf-idf (the normalized frequency of a word is divided by the logarithm of its frequency in the British National Corpus (BNC) word frequency list<ref type="foot" target="#foot_2">2</ref> ). Figures 2-5 show the results of our tests. Each figure contains three curves: the first is obtained by filtering the items without taking into account novelty (having the interest factor of each neuron always equal to 1). For the second, we used the adaptation of <ref type="bibr">(7)</ref>, with λ = 0.1 and a relaxation time long enough so that once a neuron has won, its interest factor will remain low for the duration of the test; the third is obtained using the neighborhood adaptation (9), with τ = 2. The values of λ and τ are not critical, and some preliminary tests on limited data sets allowed us to select reasonable values. Figures <ref type="figure" target="#fig_5">2 and 3</ref> show the precision and the coverage V for filtering done with simple weighting (without an idf term). The precision drops only marginally when considering novelty, but the number of neurons that "win" is increased by more than an order of magnitude, confirming that the results are relative to a broader portion of the user context.</p><p>Figures <ref type="figure" target="#fig_8">4 and 5</ref> show the same results using a weighting scheme with the idf term for the news items. The results are qualitatively the same, although we observe a general drop in precision. We have not investigated the causes of this drop, which we suspect are related to the small number of words in the items. Note that in the range 200-500 results, the system with novelty actually achieves a better precision that the simple system. We haven't thoroughly investigated this phenomenon, but it is possible that the simple system get "stuck" in groups of similar documents from other categories with a similar (and relatively high) value of measred relevance. Introducing novelty in this case would reduce the presence of these groups.   We must also remark that the MCAT category is quite difficult to filter, as many of the significant terms that appear in it are part of the economics jargon, and appear in other categories as well, such as ECAT. This explains the relatively low values of precision that we obtain, independently of use of the method for increasing novelty. This might also explain why in some cases the novelty system attains a higher precision than the one based only on relevance: there are probably parts of the context more ambiguous than others sine they are sensitive to words that appear in categories other than MCAT. With the novelty system, once an item has activated these part, they become "desensitized," and cease to produce irrelevant results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">CONCLUSION</head><p>In this paper, we have presented a method for increasing the novelty of a recommendation system that receives a stream, as opposed to the session-based recommender systems common in the literature. We have shown that a model of the user context can be useful to increase the novelty of the recommended results without having to receive results irrelevant for the user. The fact that we receive a potentially infinite stream of items poses new challenges for novelty increasing systems, as redundancy, which is taken to be a constant in session based system, "fades away" in the case of stream: a result very similar to one already received will be considered redundant in the short term, but may become interesting again in the long term. To deal with this problem, we have introduced an interest factor in our model that is reduced when a certain area of the context is activated but that, if the area is no longer activated, recovers in a time that can be controlled by design.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>6 Figure 1 :</head><label>61</label><figDesc>Figure 1: Schema of out stream recommendation system.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Precision vs. number of recommended items (simple weighting).</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Coverage vs. number of recommended items (simple weighting)</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_7"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Precision vs. number of recommended items (tf-idf weighting)</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_8"><head>Figure 5 :</head><label>5</label><figDesc>Figure 5: Coverage vs. number of recommended items (tf-idf weighting)</figDesc></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">The map can be k dimensional; however, maps with k &gt; 4 are rarely used because of the combinatorial explosion in the number of neuron which would lead to over-fitting most training sets. We have used mostly maps with k =</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">, and we shall consider only this case here, mainly because it avoid notational complexities.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_2">Available at http://www.natcorp.ox.ac.uk/</note>
		</body>
		<back>

			<div type="funding">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>* This work was supported by the Ministerio de Educación y Ciencia under the grant N. TIN2011-28538-C02, Novelty, diversity, context and time: newdimensions in next-generation information retrieval and recommender systems.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title/>
		<author>
			<persName><surname>References</surname></persName>
		</author>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Challenges in information retrieval and language modeling: report of a workshop held at the center for intelligent information retrieval</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">E A</forename><surname>Allan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">SIGIR Forum</title>
		<imprint>
			<biblScope unit="volume">37</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="31" to="47" />
			<date type="published" when="2003-04">Apr. 2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Newsjunkie: providing personalized newsfeeds via analysis of information novelty</title>
		<author>
			<persName><forename type="first">E</forename><surname>Gabrilovich</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Dumais</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Horvitz</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">WWW &apos;04: Proceedings of the 13th international conference on World Wide Web</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2004">2004</date>
			<biblScope unit="page" from="482" to="490" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Computationally efficient approximation of a probabilistic model for document representation in the WEBSOM full-text analysis method</title>
		<author>
			<persName><forename type="first">S</forename><surname>Kaski</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Neural Processing letters</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="issue">2</biblScope>
			<date type="published" when="1997">1997</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Rcv1: A new benchmark collection for text categorization research</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">D</forename><surname>Lewis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">G</forename><surname>Rose</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Li</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">J. Mach. Learn. Res</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="page" from="361" to="397" />
			<date type="published" when="2004-12">Dec. 2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">An information-pattern-based approach to novelty detection</title>
		<author>
			<persName><forename type="first">X</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">W</forename><surname>Croft</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Information Processing &amp; Management</title>
		<imprint>
			<biblScope unit="volume">44</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="1159" to="1188" />
			<date type="published" when="2008-05">May 2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Contextual search: Issues and challenges</title>
		<author>
			<persName><forename type="first">G</forename><surname>Pasi</surname></persName>
		</author>
		<idno>978 − 3 − 642 − 25364 − 53</idno>
	</analytic>
	<monogr>
		<title level="m">Information Quality in e-Health</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">A</forename><surname>Holzinger</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">K.-M</forename><surname>Simonic</surname></persName>
		</editor>
		<meeting><address><addrLine>Berlin / Heidelberg</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="1007">2011. 1007</date>
			<biblScope unit="volume">7058</biblScope>
			<biblScope unit="page">10</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Context as a non-ontological determinant of semantics</title>
		<author>
			<persName><forename type="first">S</forename><surname>Santini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Dumitrescu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 3rd International conference on Semantics and digital media technologies</title>
				<meeting>the 3rd International conference on Semantics and digital media technologies</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Identification of ambiguous queries in web search</title>
		<author>
			<persName><forename type="first">R</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Luo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J.-Y</forename><surname>Nie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H.-W</forename><surname>Hon</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Information Processing &amp; Management</title>
		<imprint>
			<biblScope unit="volume">45</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="216" to="229" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
