<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Collaborative Tag Recommendation System based on Logistic Regression</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Zhen</forename><surname>Liao</surname></persName>
							<email>liaozhen@mail.nankai.edu.cn</email>
							<affiliation key="aff8">
								<orgName type="department">College of Information Technology Science</orgName>
								<orgName type="institution">Nankai University</orgName>
								<address>
									<settlement>Tianjin</settlement>
									<country key="CN">China</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Maoqiang</forename><surname>Xie</surname></persName>
							<affiliation key="aff9">
								<orgName type="department">College of Software</orgName>
								<orgName type="institution">Nankai University</orgName>
								<address>
									<settlement>Tianjin</settlement>
									<country key="CN">China</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Hao</forename><surname>Cao</surname></persName>
							<email>caohao@mail.nankai.edu.cn</email>
							<affiliation key="aff9">
								<orgName type="department">College of Software</orgName>
								<orgName type="institution">Nankai University</orgName>
								<address>
									<settlement>Tianjin</settlement>
									<country key="CN">China</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Yalou</forename><surname>Huang</surname></persName>
							<email>huangyl@nankai.edu.cn</email>
							<affiliation key="aff9">
								<orgName type="department">College of Software</orgName>
								<orgName type="institution">Nankai University</orgName>
								<address>
									<settlement>Tianjin</settlement>
									<country key="CN">China</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Marek</forename><surname>Lipczak</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Yeming</forename><surname>Hu</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Yael</forename><surname>Kollet</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Evangelos</forename><surname>Milios</surname></persName>
						</author>
						<author>
							<persName><forename type="first">E</forename><surname>Montañés</surname></persName>
						</author>
						<author>
							<persName><forename type="first">J</forename><forename type="middle">R</forename><surname>Quevedo</surname></persName>
						</author>
						<author>
							<persName><forename type="first">I</forename><surname>Díaz</surname></persName>
						</author>
						<author>
							<persName><forename type="first">J</forename><surname>Ranilla</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Johannes</forename><surname>Mrosek</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Stefan</forename><surname>Bussmann</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Hendrik</forename><surname>Albers</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Kai</forename><surname>Posdziech</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Benedikt</forename><surname>Hengefeld</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Nils</forename><surname>Opperman</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Stefan</forename><surname>Robert</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Gerrit</forename><surname>Spira</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Hendri</forename><surname>Murfi</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Klaus</forename><surname>Obermayer</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Cataldo</forename><surname>Musto</surname></persName>
							<email>musto@di.uniba.it</email>
						</author>
						<author>
							<persName><forename type="first">Fedelucio</forename><surname>Narducci</surname></persName>
							<email>narducci@di.uniba.it</email>
						</author>
						<author>
							<persName><forename type="first">Marco</forename><surname>De Gemmis</surname></persName>
							<email>degemmis@di.uniba.it</email>
						</author>
						<author>
							<persName><forename type="first">Pasquale</forename><surname>Lops</surname></persName>
							<email>lops@di.uniba.it</email>
						</author>
						<author>
							<persName><forename type="first">Giovanni</forename><surname>Semeraro</surname></persName>
							<email>semeraro@di.uniba.it</email>
						</author>
						<author>
							<persName><forename type="first">Ilari</forename><forename type="middle">T</forename><surname>Nieminen</surname></persName>
							<email>ilari.nieminen@tkk.fi</email>
						</author>
						<author>
							<persName><forename type="first">Steffen</forename><surname>Rendle</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Lars</forename><surname>Schmidt-Thieme</surname></persName>
							<email>schmidt-thieme@ismll.uni-hildesheim.de</email>
						</author>
						<author>
							<persName><forename type="first">Xiance</forename><surname>Si</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Zhiyuan</forename><surname>Liu</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Peng</forename><surname>Li</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Qixia</forename><surname>Jiang</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Maosong</forename><surname>Sun</surname></persName>
						</author>
						<author>
							<persName><roleName>Liangjie</roleName><forename type="first">Jian</forename><surname>Wang</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Brian</forename><forename type="middle">D</forename><surname>Davison</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Robert</forename><surname>Wetzker</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Alan</forename><surname>Said</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Carsten</forename><surname>Zimmermann</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Ning</forename><surname>Zhang</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Yuan</forename><surname>Zhang</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Jie</forename><surname>Tang</surname></persName>
							<email>jietang@tsinghua.edu.cn</email>
						</author>
						<author>
							<persName><forename type="first">Leandro</forename><forename type="middle">Balby</forename><surname>Marinho</surname></persName>
							<email>marinho@ismll.uni-hildesheim.de</email>
						</author>
						<author>
							<persName><forename type="first">Christine</forename><surname>Preisach</surname></persName>
							<email>preisach@ismll.uni-hildesheim.de</email>
						</author>
						<author>
							<persName><forename type="first">Iván</forename><surname>Cantador</surname></persName>
							<email>cantador@dcs.gla.ac.uk</email>
						</author>
						<author>
							<persName><forename type="first">David</forename><surname>Vallet</surname></persName>
							<email>dvallet@dcs.gla.ac.uk</email>
						</author>
						<author>
							<persName><forename type="first">Joemon</forename><forename type="middle">M</forename><surname>Jose</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Alejandro</forename><surname>Bellogín</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Pablo</forename><surname>Castells</surname></persName>
						</author>
						<author>
							<persName><forename type="first">A</forename><forename type="middle">I</forename><surname>Communications</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Lian</forename><surname>Xue</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Chunhua</forename><surname>Liu</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Fei</forename><surname>Teng</surname></persName>
							<email>nktengfei@mail.nankai.edu.cn</email>
						</author>
						<author>
							<persName><forename type="first">Szymon</forename><surname>Chojnacki</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Jonathan</forename><surname>Gemmell</surname></persName>
							<email>jgemmell@cti.depaul.edu</email>
						</author>
						<author>
							<persName><forename type="first">Maryam</forename><surname>Ramezani</surname></persName>
							<email>mramezani@cti.depaul.edu</email>
						</author>
						<author>
							<persName><forename type="first">Thomas</forename><surname>Schimoler</surname></persName>
							<email>tschimoler@cti.depaul.edu</email>
						</author>
						<author>
							<persName><forename type="first">Laura</forename><surname>Christiansen</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Bamshad</forename><surname>Mobasher</surname></persName>
							<email>mobasher@cti.depaul.edu</email>
						</author>
						<author>
							<persName><forename type="first">Anestis</forename><surname>Gkanogiannis</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Theodore</forename><surname>Kalamboukis</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Sanghun</forename><surname>Ju</surname></persName>
							<email>shju@ml.ssu.ac.kr</email>
						</author>
						<author>
							<persName><forename type="first">Kyu-Baek</forename><surname>Hwang</surname></persName>
							<email>kbhwang@ssu.ac.kr</email>
						</author>
						<author>
							<persName><forename type="first">Thomas</forename><surname>Kleinbauer</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Sebastian</forename><surname>Germesin</surname></persName>
						</author>
						<author>
							<affiliation key="aff0">
								<orgName type="laboratory">Information Systems and Machine Learning Lab (ISMLL) Samelsonplatz 1</orgName>
								<orgName type="institution">University of Hildesheim</orgName>
								<address>
									<postCode>D-31141</postCode>
									<settlement>Hildesheim</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff1">
								<orgName type="department">Department of Computing Science</orgName>
								<orgName type="institution">University of Glasgow</orgName>
								<address>
									<addrLine>Lilybank Gardens</addrLine>
									<postCode>G12 8QQ</postCode>
									<settlement>Glasgow</settlement>
									<country key="GB">Scotland, UK</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff2">
								<orgName type="department">College of Software</orgName>
								<orgName type="institution">Nankai University</orgName>
								<address>
									<settlement>Tianjin</settlement>
									<country key="CN">P.R.China</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff3">
								<orgName type="department" key="dep1">Department of Artificial Intelligence</orgName>
								<orgName type="department" key="dep2">Institute of Computer Science</orgName>
								<orgName type="institution">Polish Academy of Sciences</orgName>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff4">
								<orgName type="department">Center for Web Intelligence School of Computing</orgName>
								<orgName type="institution">DePaul University Chicago</orgName>
								<address>
									<region>Illinois</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff5">
								<orgName type="department">Department of Informatics</orgName>
								<orgName type="institution" key="instit1">Athens</orgName>
								<orgName type="institution" key="instit2">University of Economics and Business</orgName>
								<address>
									<settlement>Athens</settlement>
									<country key="GR">Greece</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff6">
								<orgName type="department">School of Computing</orgName>
								<orgName type="institution">Soongsil University</orgName>
								<address>
									<postCode>156-743</postCode>
									<settlement>Seoul</settlement>
									<country key="KR">Korea</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff7">
								<orgName type="department">German Research Center for Artificial Intelligence (DFKI</orgName>
								<address>
									<postCode>66123</postCode>
									<settlement>Saarbrücken</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff10">
								<orgName type="department">Department of Computer Science</orgName>
								<orgName type="institution">University of Bari &quot;Aldo Moro&quot;</orgName>
								<address>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff11">
								<orgName type="institution">Helsinki University of Technology</orgName>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff12">
								<orgName type="department">Knowledge Engineering Group Department of Computer Science and Technology</orgName>
								<orgName type="institution">Tsinghua University</orgName>
								<address>
									<settlement>Beijing</settlement>
									<country key="CN">China</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff13">
								<orgName type="department">Knowledge Engineering Group Department of Computer Science and Technology</orgName>
								<orgName type="institution">Tsinghua University</orgName>
								<address>
									<settlement>Beijing</settlement>
									<country key="CN">China</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Collaborative Tag Recommendation System based on Logistic Regression</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">37B8827F2E9AE30715B21558A2C927A1</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T06:53+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>social tag recommendation</term>
					<term>co-occurrence</term>
					<term>graph vertex centrality</term>
					<term>collaborative bookmarking multilayer</term>
					<term>ontology</term>
					<term>hybrid</term>
					<term>recommendation</term>
					<term>configwork</term>
					<term>aicom</term>
					<term>ai</term>
					<term>communication</term>
					<term>user</term>
					<term>preference</term>
					<term>semantic</term>
					<term>concept</term>
					<term>domain</term>
					<term>ontology</term>
					<term>item</term>
					<term>space</term>
					<term>way</term>
					<term>cluster</term>
					<term>similarity</term>
					<term>individual</term>
					<term>layer</term>
					<term>community</term>
					<term>interest Information Retrieval</term>
					<term>Searching Engines</term>
					<term>Tag Recommendations Folksonomies</term>
					<term>Tag Recommenders</term>
					<term>Hybrid Recommenders social bookmarking</term>
					<term>folksonomy</term>
					<term>tag recommendation example new example de bibtex Dict/Split: example new example bibtex coloradoboomerangs Dict/Split: colorado boomerangs Recommender system</term>
					<term>Social tagging</term>
					<term>Machine learning</term>
					<term>Keyword extraction</term>
					<term>Concept extraction Recommender Systems</term>
					<term>Web 2.0</term>
					<term>Collaborative Tagging Systems</term>
					<term>Folksonomies</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Folksonomy data is relational by nature, and therefore methods that directly exploit these relations are prominent for the tag recommendation problem. Relational methods have been successfully applied to areas in which entities are linked in an explicit manner, like hypertext documents and scientific publications. For approaching the graph-based tag recommendation task of the ECML PKDD Discovery Challenge 2009, we propose to turn the folksonomy graph into a homogeneous post graph and use relational classification techniques for predicting tags. Our approach features adherence to multiple kinds of relations, semi-supervised learning and fast predictions.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Preface</head><p>Since 1999 the ECML PKDD embraces the tradition of organizing a Discovery Challenge, allowing researchers to develop and test algorithms for novel and real world datasets. This year's Discovery Challenge<ref type="foot" target="#foot_0">1</ref> presents a dataset from the field of social bookmarking to deal with the recommendation of tags. The results submitted by the challenge's participants are presented at an ECML PKDD workshop on September 7th, 2009, in Bled, Slovenia.</p><p>The provided dataset has been created using data of the social bookmark and publication sharing system BibSonomy,<ref type="foot" target="#foot_1">2</ref> operated by the organizers of the challenge. The training data was released on March 25th 2009, the test data on July 6th. The participants had time until July 8th to submit their results. This gave researchers 14 weeks time to tune their algorithms on a snapshot of a real world folksonomy dataset and 48 hours to compute results on the test data.</p><p>To support the user during the tagging process and to facilitate the tagging, BibSonomy includes a tag recommender. When a user finds an interesting web page (or publication) and posts it to BibSonomy, the system offers up to five recommended tags on the posting page. The goal of the challenge is to learn a model which effectively predicts the keywords a user has in mind when describing a web page (or publication). We divided the problem into three tasks, each of which focusing on a certain aspect. All three tasks get the same dataset for training. It is a snapshot of BibSonomy until December 31st 2008. The dataset is cleaned and consists of two parts, the core part and the complete snapshot. The test dataset is different for each task.</p><p>Task 1: Content-Based Tag Recommendations. The test data for this task contains posts, whose user, resource or tags are not contained in the post-core at level 2 of the training data. Thus, methods which can't produce tag recommendations for new resources or are unable to suggest new tags very probably won't produce good results here.</p><p>Task 2: Graph-Based Recommendations. This task is especially intended for methods relying on the graph structure of the training data only. The user, resource, and tags of each post in the test data are all contained in the training data's post-core at level 2.</p><p>Task 3: Online Tag Recommendations. This is a bonus task which will take place after Tasks 1 and 2. The participants shall implement a recommendation service which can be called via HTTP by BibSonomy's recommender infrastructure when a user posts a bookmark or publication. All participating recommenders are called on each posting process, one of them is chosen to actually deliver the results to the user. We can then measure the performance of the recommenders in an online setting, where timeouts are important and where we can measure which tags the user has clicked on.</p><p>Results. More than 150 participants registered for the mailing list which enabled them to download the dataset. At the end, we received 42 submissions -21 for each of the Tasks 1 &amp; 2. Additionally, 24 participants submitted a paper that can be found in the proceedings at hand. We used the F1-Measure common in Information Retrieval to evaluate the submitted recommendations. Therefore, we first computed for each post in the test data precision and recall by comparing the first five recommended tags against the tags the user has originally assigned to this post. Then we averaged precision and recall over all posts in the test data and used the resulting precision and recall to compute the F1-Measure as f1m = 2•precision • recall precision + recall . The winning team of Task 1 has an f1m of 0.18740, the second and third follow with 0.18001 and 0.17975. For Task 2, the winner achieved an f1m of 0.35594, followed by 0.33185 and 0.32461. The winner of Task 3 will be announced at the conference and later on the website of the challenge.</p><p>Lipczak et al. from Dalhousie University, Halifax, Canada (cf. page 157) are the winners of Task 1. With a method based on the combination of tags from the resource's title, tags assigned to the resource by other users and tags in the user's profile they reached an f1m of 0.18740 in Task 1 and additionally achieved the third place in Task 2 with an f1m of 0.32461. The system is composed of six recommenders and the basic idea is to augment the tags from the title by related tags extracted from two tag-tag-co-occurrence graphs and from the user's profile and then rescore and merge them.</p><p>The winners of Task 2, Rendle and Schmidt-Thieme from University of Hildesheim, Germany (cf. page 235) achieved an f1m of 0.35594 with a statistical method based on factor models. Therefore, they factorize the folksonomy structure to find latent interactions between users, resources and tags. Using a variant of the stochastic gradient descent algorithm the authors optimize an adaptation of the Bayesian Personal Ranking criterion. Finally, they estimate how many tags to recommend to further improve precision.</p><p>The second of Task 1 (Mrosek et al., page 189) harvests tags from sources like Delicious, Google Scholar, and CiteULike. They also employ the full-text of web pages and PDFs. The third (Ju and Hwang, page 109) merges tags which have been earlier assigned to the resource or used by the user as well as resource descriptions by a weighting scheme. Finally, the second of Task 2 (Balby Marinho et al., page 7) uses relational classification methods in a semi-supervised learning scenario to recommend tags.</p><p>We thank all participants of the challenge for their contributions and the organizers of the ECML PKDD 2009 conference for their support. Furthermore, we want to thank our sponsors Nokia <ref type="foot" target="#foot_2">3</ref> and Tagora <ref type="foot" target="#foot_3">4</ref> for supporting the challenge by awarding prizes for the winners of each task. We are looking forward to a very exciting and interesting workshop.</p><p>Kassel, August 2009 Folke Eisterlehner, Andreas Hotho, Robert Jäschke</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>One might want tag recommendations for several reasons, as for example: simplifying the tagging process for the user, exposing different facets of a resource and helping the tag vocabulary to converge. Given that users are free to tag, i.e., the same resource can be tagged differently by different people, it is important to personalize the recommended tags for an individual user.</p><p>Tagging data forms a ternary relation between users, resources and tags, differently from typical recommender systems in which the relation is usually binary between users and resources. The best methods presented so far explore this ternary relation to compute tag predictions, either by means of tensor factorization <ref type="bibr">[8]</ref> or PageRank <ref type="bibr">[3]</ref>, on the hypergraph induced by the ternary relational data. We, on the other hand, propose to explore the underlying relational graph between posts by means of relational classification.</p><p>In this paper we describe our approaches for addressing the graph-based tag recommendation task of the ECML PKDD Discovery Challenge 2009. We present two basic algorithms: PWA* (probabilistic weighted average), an iterative relational classification algorithm enhanced with relaxation labelling, and WA* (weighted average), an iterative relational classification method without relaxation labelling. These methods feature: adherence to multiple kinds of relations, training free, fast predictions, and semi-supervised classification. Semi-supervised classification is particularly appealing because it allows us to evtl. benefit from the information contained in the test dataset. Furthermore, we propose to combine these methods through unweighted voting.</p><p>The paper is organized as follows. Section 2 presents the notation used throughout the paper. In Section 3 we show how we turned the folksonomy into a post relational graph. Section 4 introduces the individual classifiers and the ensemble technique we used. In Section 5 we elaborate on the evaluation and experiments conducted for tuning the parameters of our models, and report the results obtained on the test dataset released for the challenge. The paper closes with conclusions and directions for future work.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Notation</head><p>Foksonomy data usually comprises a set of users U , a set of resources R, a set of tags T , and a set Y of ternary relations between them, i.e., Y ⊆ U × R × T .</p><p>Let X := {(u, r) | ∃t ∈ T : (u, r, t) ∈ Y } be the set of all unique user/resources combinations in the data, where each pair is called a post. For convenience, let T (x = (u, r)) := {t ∈ T | (u, r, t) ∈ Y } be the set of all tags assigned to a given post x ∈ X. We consider train/test splits based on posts, i.e., X train , X test ⊂ X disjoint and covering all of X:</p><formula xml:id="formula_0">X train ∪X test = X</formula><p>For training, the learner has access to the set X train of training posts and their true tags T | Xtrain . The tag recommendation task is then to predict, for a given x ∈ X test , a set T (x) ⊆ T of tags that are most likely to be used by the resp. user for the resp. resource.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Relation Engineering</head><p>We propose to represent folksonomy data as a homogeneous, undirected relational graph over the post set, i.e., G := (X, E) in which edges are annotated with a weight w : X × X → R denoting the strength of the relation. Besides making the input data more compact -we have only a binary relation R ⊆ X ×X between objects of the same type -this representation will allow us to trivially cast the problem of personalized tag recommendations as a relational classification problem.</p><p>Relational classifiers usually consider, additionally to the typical attribute-value data of objects, relational information. A scientific paper, for example, can be connected to another paper that has been written by the same author or because they share common citations. It has been shown in many classification problems that relational classifiers perform better than purely attribute-based classifiers <ref type="bibr">[1,</ref><ref type="bibr">4,</ref><ref type="bibr">6]</ref>.</p><p>In our case, we assume that posts are related to each other if they share the same user: R user := {(x, x ) ∈ X × X | user(x) = user(x )}, the same resource: R res := {(x, x ) ∈ X × X|res(x) = res(x )}, or either share the same user or resource: R res user := R user ∪ R res (see Figure <ref type="figure" target="#fig_0">1</ref>). For convenience, let user(x) and res(x) denote the user and resource of post x respectively. Thus, each post is connected to each other either in terms of other users that tagged the same resource, or the resources tagged by the same user. Weights are discussed in Section 4. Note that it may happen that some of the related posts belong themselves to the test dataset, allowing us to evtl. profit from the unlabeled information of test nodes through, e.g., collective inference (see <ref type="bibr">Section 4)</ref>. Thus, differently from other approaches (e.g., <ref type="bibr">[3,</ref><ref type="bibr">8]</ref>) that are only restricted to X train , we can also exploit the set X test of test posts, but of course not their associated true tags. Now, for a given x ∈ X test , one can use the tagging information of related instances to estimate T (x). A simple way to do that is, e.g., through tag frequencies of related posts:</p><formula xml:id="formula_1">P (t|x) := |{x ∈ N x |t ∈ T (x )}| |N x | , x ∈ X, t ∈ T<label>(1)</label></formula><p>while N x is the neighborhood of x:</p><formula xml:id="formula_2">N x := {x ∈ X | (x, x ) ∈ R, T (x) = ∅}<label>(2)</label></formula><p>In section 4 we will present the actual relational classifiers we have used to approach the challenge.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Relational Classification for Tag Recommendation</head><p>We extract the relational information by adapting simple statistical relational methods, usually used for classification of hypertext documents, scientific publications or movies, to the tag recommendation scenario. The aim is to recommend tags to users by using the neighborhood encoded in the homogeneous graph G(X, E). Therefore we described a very simple method in eq. ( <ref type="formula" target="#formula_1">1</ref>), where the probability for a tag t ∈ T given a node x (post) is computed by counting the frequency of neighboring posts x ∈ N x that have used the same tag t. In this case the strength of the relations is not taken into account, i.e., all considered neighbors of x have the same influence on the probability of tag t given x. But this is not an optimal solution, the more similar posts are to each other the higher the weight of this edge should be.</p><p>Hence, a more suitable relational method for tag recommendation is the WeightedAverage (WA) which sums up all the weights of posts x ∈ N x that share the same tag t ∈ T and normalizes this by the sum over all weights in the neighborhood.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>P (t|x) =</head><p>x ∈Nx|t∈T (x ) w(x, x )</p><p>x ∈Nx w(x, x )</p><p>Thus, WA does only consider neighbors that belong to the training set.</p><p>A more sophisticated relational method that takes probabilities into account is the probabilistic weighted average (PWA), it calculates the probability of t given x by building the weighted average of the tag probabilities of neighbor nodes x ∈ N x :</p><formula xml:id="formula_4">P (t|x) =</formula><p>x ∈Nx w(x, x )P (t|x )</p><p>x ∈Nx w(x, x )</p><p>Where P (t|x ) = 1 for x ∈ X train , i.e., we are only exploiting nodes contained in the training set (see eq. ( <ref type="formula" target="#formula_2">2</ref>)). We will see in the next paragraph how one can exploit these probabilities in a more clever way. Both approaches have been introduced in <ref type="bibr">[5]</ref> and applied to relational datasets.</p><p>Since we want to recommend more than one tag we need to cast the tag recommendation problem as a multilabel classification problem, i.e., assign one or more classes to a test node. We accomplish the multilabel problem by sorting the calculated probabilities P (t|x) for all x ∈ X test and recommend the top n tags with highest probabilities.</p><p>The proposed relational methods could either be applied on R res user , i.e., the union of the user and resource relation or on each relation R user , R res individually. If applied on each relation the results could be combined by using ensemble techniques.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Semi-Supervised Learning</head><p>As mentioned before, we would like additionally, to exploit unlabeled information contained in the graph and use the test nodes that have not been tagged yet, but are related to other nodes. This can be achieved by applying collective inference methods, being iterative procedures, which classify related nodes simultaneously and exploit relational autocorrelation and unlabeled data. Relational autocorrelation is the correlation among a variable of an entity to the same variable (here the class) of a related entity, i.e., connected entities are likely to have the same classes assigned. Collective Classification is semi-supervised by nature, since one exploits the unlabeled part of the data. One of this semi-supervised methods is relaxation labeling <ref type="bibr">[1]</ref>, it can be formally expressed as:</p><formula xml:id="formula_6">P (t|x) (i+1) = M (P (t|x ) (i) x ∈Nx )<label>(5)</label></formula><p>We first initialize the unlabeled nodes with the prior probability calculated using the train set, then compute the probability of tag t given x iteratively using a relational classification method M based on the neighborhood N x in the inner loop. The procedure stops when the algorithm converges (i.e., the difference of the tag probability between iteration i and i + 1 is less than a very small ) or a certain number of iterations is reached.</p><p>We used eq. ( <ref type="formula" target="#formula_5">4</ref>) as relational method inside the loop, where we do not require that the neighbors x are in the training set, but are using the probabilities of unlabeled nodes. For PWA this means that in each iteration we use the probabilities of the neighborhood estimated in the previous iteration collectively. PWA combined with collective inference is denoted as PWA* in the following.</p><p>For WeightedAverage we did not use relaxation labeling but applied a so called oneshot-estimation <ref type="bibr">[5,</ref><ref type="bibr">7]</ref>. We did only use the neighbors with known classes, i.e., in the first iteration we exploit only nodes from the training set, while in the next iteration we used also test nodes that have been classified in the previous iterations. The procedure stops when all test nodes could be classified or a specific number of iterations is reached. Hence, the tag probabilities are not being re-estimated like for the relaxation labeling but only estimated once. Thus, WA combined with the one-shot-estimation procedure is denoted as WA*.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Ensemble</head><p>Ensemble classification may lead to significant improvement on classification accuracy, since uncorrelated errors made by the individual classifiers are removed by the combination of different classifiers <ref type="bibr">[2,</ref><ref type="bibr">6]</ref>. Furthermore, ensemble classification reduces variance and bias.</p><p>We have decided to combine WA* and PWA* through a simple unweighted voting, since voting performs particularly well when the results of individual classifiers are similar; as we will see in Section 5, WA* and PWA* yielded very similar results in our holdout set.</p><p>After performing the individual classifiers, we receive probability distributions for each classifier K l as output and build the arithmetic mean of the tag-assignment probabilities for each test post and tag:</p><formula xml:id="formula_7">P (t|x) = 1 L • L l=1 P l (t|x), L := |K l |P l (t|x) = 0, t ∈ T |<label>(6)</label></formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">Weighting Schemes</head><p>The weight w in eq. ( <ref type="formula" target="#formula_3">3</ref>) and ( <ref type="formula" target="#formula_5">4</ref>) is an important factor in the estimation of tag probabilities, since it describes the strength of the relation between x and x . There are several ways to estimate these weights:</p><p>1. For two nodes (x, x ) ∈ R res , compute their similarity by representing x and x as user-tag profile vectors. Each component of the profile vector corresponds to the count of co-occurrences between users and tags:</p><formula xml:id="formula_8">φ user-tag := (|Y ∩ ({user(x)} × R × {t})|) t∈T 2.</formula><p>Similarly to 1, for two nodes (x, x ) ∈ R user , the node similarity is computed by representing x and x as resource-tag profile vectors: The edge weight is finally computed by applying the cosine similarity over the desired profile vectors: sim(φ(x), φ(x )) := φ(x), φ(x ) φ(x) φ(x )</p><p>In our experiments we basically used the scheme 1, since there is no new user in the data and therefore we can always build user-tag profile vectors.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Evaluation</head><p>All the results presented in this section are reported in terms of F1-score, the official measure used by the graph-based tag recommendation task of the ECML PKDD Discovery Challenge 2009. For a given x ∈ X test the F1-Score is computed as follows:</p><formula xml:id="formula_10">F1-score T (x) = 2 • Recall T (x) • Precision T (x)</formula><p>Recall T (x) + Precision T (x) <ref type="bibr">(8)</ref> Although the methods presented in Section 4 usually do not have free parameters, we realized that R user and R res can have a different impact in the recommendation quality (cf. Figures <ref type="figure" target="#fig_52">2 and 3</ref>), and thereby we introduced a parameter to reward the best relations in R res user by a factor c ∈ N: if R res yields better recommendations than R user for example, all edge weights in R res user that refer to R res are multiplied by c. For searching the best c value we performed a greedy search over the factor range {1, ..., 4} on a holdout set created by randomly selecting 800 posts from the training data. Tables <ref type="table" target="#tab_120">1 and 2</ref> show the characteristics of the training and test/holdout datasets respectively. Figure <ref type="figure" target="#fig_4">2</ref> presents the results of WA*-Full<ref type="foot" target="#foot_4">1</ref> , i.e., WA* performed over R res user , for different c values on the holdout set according to the F1-score. We also plot the results of WA*-Res and WA*-Usr (i.e., WA* on R res and R user resp.).</p><p>After finding the best c value on the holdout set, we applied WA*-Full, PWA*-Full, and the ensemble (c.f. eq. 6) to the challenge test dataset (see Figure <ref type="figure" target="#fig_10">3</ref>). Note that the results on the challenge test dataset are much lower than those on the holdout set. It may indicate that either our holdout set was not a good representative of the population or that the challenge test dataset represents a concept drift. We plan to further investigate the reasons underlying this large deviation.</p><p>According to the rules of the challenge, the F1-score is measured over the Top-5 recommended tags, even though one is not forced to always recommend 5 tags. This is an important remark because if one recommends more tags than the true number of tags attached to a particular test post, one can lower precision. Therefore, we estimate the number of tags to be recommended to each test post by taking the average number of tags used by each test user to his resources. If a given test user has tagged his resources with 3 tags in average, for example, we recommend the Top-3 tags delivered by our algorithms for all test posts in which this user appears.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Conclusions</head><p>In this paper we proposed to approach the graph-based tag recommendation task of the ECML PKDD Discovery Challenge 2009 by means of relational classification. We first turned the usual tripartite graph of social tagging systems into a homogeneous post graph, whereby simple statistical relational methods can be easily applied. Our methods are training free and the prediction runtime only depends on the number of neighbors and tags, which is fast since the training data is sparse. The models we presented also incorporate a semi-supervised component that can evtl. benefit from test data. We presented two relational classification methods, namely WA* and PWA*, and one ensemble based on unweighted voting over the tag probabilities delivered by these methods.</p><p>We also introduced a parameter in order to reward more informative relations, which was learned through a greedy search in a holdout set.</p><p>In future work we want to investigate new kinds of relations between the posts (e.g. content-based), other ensemble techniques, and new methods for automatically learning more informative weights. Learning optimal ranking with tensor factorization for tag recommendation. In KDD '09: Proceedings of the 15th ACM SIGKDD international conference on Knowledge discovery and data mining, pages 727-736. ACM, 2009.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Social tagging systems allow users to create or upload resources (web pages <ref type="foot" target="#foot_5">1</ref> , scientific publications<ref type="foot" target="#foot_6">2</ref> , photos <ref type="foot" target="#foot_7">3</ref> , video clips <ref type="foot" target="#foot_8">4</ref> , music tracks <ref type="foot" target="#foot_9">5</ref> ), annotate them with freely chosen words -so called tags -and share them with others. The set of users, resources, tags and annotations (i.e., triplets user-tag-resource) is commonly known as folksonomy, and constitutes a collective unstructured knowledge classification. This implicit classification is then used by users to organise, explore and search for resources, and by systems to recommend users interesting resources. These systems usually include tag recommendation mechanisms to ease the finding of relevant tags for a resource, and consolidate the tag vocabulary across users. However, as stated in <ref type="bibr">[7]</ref>, no algorithmic details have been published, and it is assumed that, in general, tag recommendations in current applications are based on suggesting those tags that most frequently were assigned to the resource, or to similar resources.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Content-based tag recommenders</head><p>Mishne <ref type="bibr">[10]</ref> presents a simple content-based tag recommender. Once a user supplies a new bookmark, bookmarks that are similar to it are identified. The tags assigned to these bookmarks are aggregated, creating a ranked list of likely tags. Then, the system filters and re-ranks the tag list. The top ranked tags are finally suggested to the user. To find similar bookmarks, the author utilises a document index, and keywords of the input bookmark to form a query that is launched against the index. The tags are scored according to their frequencies in the top results of the above query, and those tags that have been used previously by the user are boosted by a constant factor. Our approach follows the same stages, also using an index to retrieve similar bookmarks. It includes, however, more sophisticated methods of tag ranking based on tag popularity and personalisation aspects.</p><p>Byde et al. <ref type="bibr">[3]</ref> present a personalised tag recommendation method on the basis of similarity metrics between a new document and documents previously tagged by the user. These metrics are derived either from tagging data, or from content analysis, and are based on the cosine similarity metric <ref type="bibr">[14]</ref>. Similar metrics are used by our approach in some of its stages.</p><p>Chirita et al. <ref type="bibr">[4]</ref> suggest a method called P-TAG that automatically generates personalised tags for web pages. Given a particular web page, P-TAG produces keywords relevant both to the page contents and data residing on the user's desktop, thus expressing a personalised viewpoint. A number of techniques to extract keywords from textual contents, and several metrics to compare web pages and desktop documents, are investigated. Our approach applies natural language processing techniques to extract keywords from bookmark attributes, but it can be enriched with techniques like <ref type="bibr">[4]</ref> to also analyse and exploit the textual contents of the bookmarked documents.</p><p>Tatu et al. <ref type="bibr">[16]</ref> propose to extract important concepts from the textual metadata associated to bookmarks, and use semantic analysis to generate normalised versions of the concepts. For instance, European Union, EU and European Community would be normalised to the concept european_union. Then, users and resources are represented in terms of the created conceptual space, and personalised tag recommendations are based on intersections between such representations. In our approach, synonym relations and lexical derivations between tags are implicitly taking into consideration through the exploitation of tag cooccurrence graphs.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Collaborative tag recommenders</head><p>Xu et al. <ref type="bibr">[17]</ref> propose a collaborative tag recommender that favours tags used by a large number of users on the target resource (high authority in the HITS algorithm <ref type="bibr">[8]</ref>), and minimises the overlap of concepts among the recommended tags to allow for high coverage of multiple facets. Our approach also attempts to take into account tag popularity and diversity in the recommendations through the consideration of vertex centralities in the tag co-occurrence graph.</p><p>Hotho et al. <ref type="bibr">[6]</ref> present a graph-based tag recommendation approach called FolkRank, which is an adaptation of PageRank algorithm <ref type="bibr">[12]</ref>, and is applied in the folksonomy user-resource-tag graph. Its basis is the idea that a resource tagged with important tags by important users becomes important itself. The same holds, symmetrically, for users and tags. Having a graph whose vertices are associated to users, resources and tags, the algorithm reinforces each of them by spreading their weights through the graph edges. In this work, we restrict our study to the original folksonomy graph. As a future research goal, PageRank, HITS or other graph based techniques could be applied to enhance the identification of tags with high graph centrality values.</p><p>Jäscke et al. <ref type="bibr">[7]</ref> evaluate and compare several tag recommendation algorithms: an adaptation of user-based collaborative filtering <ref type="bibr">[13]</ref>, FolkRank strategy <ref type="bibr">[6]</ref>, and methods that are based on counting tag co-occurrences. The authors show that graphbased and collaborative filtering approaches provide better results that nonpersonalised methods, and state that methods based on counting co-occurrences have low computational costs, thus being preferable for real time scenarios. Our approach is computationally cheap because it is based on a simple analysis of tag co-occurrence graphs, and includes a personalisation stage to better adjust the tag recommendations to the user's profile.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3">Hybrid tag recommenders</head><p>Heymann et al. <ref type="bibr">[5]</ref> present a technique that predicts tags for a website based on page text, anchor text, surrounding hosts, and other tags assigned to the website by users. The tag predictions are based on association rules, which, as stated by the authors, may serve as a way to link disparate vocabularies among users, and may indicate synonym and polysemy cases. As a hybrid approach, our tag recommender makes use of content-based and collaborative tag information. Nonetheless, we simplify the process limiting it to the exploitation of meta-information of the contents available in the bookmarks.</p><p>Song et al. <ref type="bibr">[15]</ref> suggest a tag recommendation method that combines clustering and mixture models. Tagged documents are represented as a triplet (words, documents, tags) by two bipartite graphs. These graphs are clustered into topics by a spectral recursive embedding technique <ref type="bibr">[18]</ref>. The sparsity of the obtained clusters is dealt with a two-way Poisson mixture model <ref type="bibr">[9]</ref>, which groups documents into components and clusters words. Inference for new documents is based on the posterior probability of topic distributions, and tags recommendations are given according to the within-cluster tag rankings.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Document and index models</head><p>To suggest tags for an input bookmark, our recommender exploits meta-information associated to it. The text contents of bookmarked documents (web pages or scientific publications) could be also taken into account, but we decided to firstly study how accurate tag recommendations can be by only using bookmarking meta-information.</p><p>In this work, we test our approach with a dataset obtained from BibSonomy system, whose bookmarks have, among others, the attributes shown in Table <ref type="table" target="#tab_0">1</ref>. The continued increase in Web usage, in particular participation in folksonomies, reveals a trend towards a more dynamic and interactive Web where individuals can organise and share resources. Tagging has emerged as the de-facto standard for the organisation of such resources, providing a versatile and reactive knowledge management mechanism that users find easy to use and understand. It is common nowadays for users to have multiple profiles in various folksonomies, thus distributing their tagging activities. In this paper, we present a method for the automatic consolidation of user profiles across two popular social networking sites, and subsequent semantic modelling of their interests utilising Wikipedia as a multi-domain model. We evaluate how much can be learned from such sites, and in which domains the knowledge acquired is focussed. Results show that far richer interest profiles can be generated for users when multiple tag-clouds are combined.</p><p>In our approach, for each bookmark, using a set of NLP tools <ref type="bibr">[2]</ref>, the text attributes title, URL, abstract and description, and extended description are processed and transformed into a weighted list of keywords. These simplified bookmark representations are then stored into an index, which will allow fast searches for bookmarks that satisfy keyword-and tag-based queries. In our implementation, we used Lucene <ref type="foot" target="#foot_12">8</ref> , which allowed us to apply keyword stemming, stop words removal, and term TF-IDF weighting.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Social tag recommendation</head><p>In this section, we describe our approach to recommend social tags for a bookmark, which does not need to be already tagged. The recommendation process is divided in 5 stages, depicted in Figure <ref type="figure" target="#fig_0">1</ref>. Each of these stages is explained in detail in the next subsections. For a better understanding, the explanations follow a common illustrative example. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Extracting bookmark keywords</head><p>The first stage of our tag recommendation approach (identified by label 1 in Figure <ref type="figure" target="#fig_0">1</ref>) is the extraction of keywords from some of the textual contents of the input bookmark.</p><p>According to the document model explained in Section 2, we extract such keywords from the title, URL, abstract, description and extended description of the Table <ref type="table" target="#tab_1">2</ref> shows the content of an example bookmark whose tag recommendations are going to be explained in the rest of this section. It also lists the keywords extracted from the bookmark in the first stage of our approach. The bookmarked document is a scientific publication. Its main research fields are recommender systems and semantic web technologies. It describes a content-based collaborative recommendation model that exploits semantic (ontology-based) descriptions of user and item profiles.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Searching for similar bookmarks</head><p>The second stage (label 2 in Figure <ref type="figure" target="#fig_0">1</ref>) consists of searching for bookmarks that contain some of the keywords obtained in the previous stage.</p><p>The list of keywords extracted from the input bookmark are weighted based on their appearance frequency in the bookmark attributes, and are included in a weighted keyword-based query. This query represents an initial description of the input bookmark.</p><p>More specifically, in the query ‫ݍ‬ for bookmark ܾ , the weight ‫ݍ‬ , ∈ [0,1] assigned to each keyword ݇ is computed as the number of times the keyword appears in the bookmark attributes divided by the total number of keywords extracted from the bookmark:</p><formula xml:id="formula_11">‫ܙ‬ ‫ܖ‬ = ‫ݍ‬ሺܾ ሻ = ‫ݍ{‬ ,ଵ , … , ‫ݍ‬ , , … , ‫ݍ‬ , }</formula><p>where</p><formula xml:id="formula_12">‫ݍ‬ , = ݂ , ∑ ݂ , ୀ ୀଵ</formula><p>, being ݂ , the number of times keyword ݇ appears in bookmark ܾ fields.</p><p>The query is then launched against the index described in Section 2. Thus, we are not only taking into account the relevance of the keywords for the input bookmark, but also ranking the list of retrieved similar bookmarks. The searching result is a set of bookmarks that are similar to the input bookmark, assuming that "similar" bookmarks have common keywords. Using the cosine similarity measure for the vector space model <ref type="bibr">[14]</ref>, the retrieved bookmarks are assigned scores ‫ݓ‬ , ∈ [0,1] that measure the similarity between the query ‫ݍ‬ (i.e., the input bookmark ܾ ) and the retrieved bookmarks ܾ :</p><formula xml:id="formula_13">‫ݓ‬ , = ‫ݍ‪݅݉ሺ‬ݏ‬ , ܾ ሻ = cosሺ‫ܙ‬ ‫ܖ‬ , ‫܊‬ ܑ ሻ = ‫ܙ‬ ‫ܖ‬ • ‫܊‬ ܑ ԡ‫ܙ‬ ‫ܖ‬ ԡԡ‫܊‬ ܑ ԡ</formula><p>For the example input bookmark, Table <ref type="table" target="#tab_3">3</ref> shows the keywords, query, and some similar bookmarks obtained in the second stage of our tag recommendation model. In this stage, we attempted to define and contextualise the vocabulary that is likely to describe the contents of the bookmarked document. For that purpose, the initial set of keywords extracted from the input bookmark was used to find related bookmarks, assuming that the keywords and social tags of the latter are useful to describe the content topics of the former.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">Obtaining related social tags</head><p>Once the set of similar bookmarks has been retrieved, in the third stage (label 3 in Figure <ref type="figure" target="#fig_0">1</ref>), we collect and weight all their social tags.</p><p>The weight assigned to each tag represents how much it contributes to the definition of the vocabulary that describes the input bookmark. Based on the scores ‫ݓ‬ , of the bookmarks retrieved in the previous stage, the weight ‫ݒ‬ of a tag ‫ݐ‬ for the input bookmark ܾ is given by:</p><formula xml:id="formula_14">‫ݒ‬ ሺ‫ݐ‬ሻ = ∑ ‫ݓ‬ , :௧ ∈ ୲ୟୱሺ ሻ .</formula><p>At this point, we could finish the recommendation process suggesting those social tags with highest weights ‫ݒ‬ . However, doing this, we are not taking into account tag popularities and tag correlations, very important features of any collaborative tagging system. In fact, we conducted experiments evaluating recommendations based on the highest weighted tags, and we obtained worse results that the ones provided by the whole approach presented herein.</p><p>Table <ref type="table" target="#tab_5">4</ref> shows a subset of the tags retrieved from the bookmarks that were retrieved in Stage 2 for the example input bookmark. The weights ‫ݒ‬ for each tag are also given in the table. In this stage, we collected the social tags that are potentially relevant for describing the input bookmarked document based on a set of related bookmarks. We assigned a weight to each tag capturing the strength of its contribution to the bookmark description. However, we realised that this measure is not enough for tag recommendation purposes, and global metrics regarding the folksonomy graph, such as tag popularities and tag correlations, have to be taken into consideration.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4">Building the global social tag co-occurrence sub-graph</head><p>In the fourth stage (label 4 in Figure <ref type="figure" target="#fig_0">1</ref>), we interconnect the social tags obtained in the previous stage through the co-occurrence values of each pair of tags.</p><p>The co-occurrence of two tags ‫ݐ‬ and ‫ݐ‬ is usually defined in terms of the number of resources (bookmarks) that have been tagged with both ‫ݐ‬ and ‫ݐ‬ . In this work, we make use of the asymmetric co-occurrence metric:</p><formula xml:id="formula_15">‫ݐ‪൫‬ܿ‬ , ‫ݐ‬ ൯ = #{݊: ‫ݐ‬ ߳ tagsሺܾ ሻ ^ ‫ݐ‬ ߳ tagsሺܾ ሻ} #{݊: ‫ݐ‬ ߳ tagsሺܾ ሻ} ,</formula><p>which assigns different values for ‫ݐ‪൫‬ܿ‬ , ‫ݐ‬ ൯ and ‫ݐ‪൫‬ܿ‬ , ‫ݐ‬ ൯ dividing the number of resources tagged with the two tags by the number of resources tagged with one of them.</p><p>Computing the co-occurrence values for each pair of tags existing in a training dataset, we build a global graph where the vertices correspond to the available tags, and the edges link tags that co-occur within at least one resource. This graph is directed and weighted: each pair of co-occurring tags is linked by two edges whose weights are the asymmetric co-occurrence values of the tags.</p><p>We propose to exploit this global graph to interconnect the tags obtained in the previous stage, and extract the ones that are more related with the input bookmark. Specifically, we create a sub-graph where the vertices are the above tags, and the edges are the same as these tags have in the global co-occurrence graph. From this sub-graph, we remove those edges whose co-occurrence values ‫ݐ‪൫‬ܿ‬ , ‫ݐ‬ ൯ are lower than the average co-occurrence value of the sub-graph vertices:</p><formula xml:id="formula_16">‫ܾ‪ሺ‬ܿ_݃ݒܽ‬ ሻ = ∑ ‫ݐ‪ሺ‬ܿ‬ , ‫ݐ‬ ሻ , #{ሺ݅, ݆ሻ: ‫ݐ‪൫‬ܿ‬ , ‫ݐ‬ ൯ &gt; 0} ,</formula><p>where ‫ݐ‬ and ‫ݐ‬ are the pairs of social tags related to the input bookmark ܾ . Removing these edges, we aim to isolate (and later discard) "noise" tags that less frequently appear in bookmark annotations. We hypothesise that vertices of the generated sub-graph that are most "strongly" connected with the rest of the vertices correspond to tags that should be recommended, assuming that high graph vertex centralities are associated to the most informative or representative vertices. In this context, it is important to note that related tags with high weights ‫ݒ‬ do not necessarily have to be the ones with highest vertex centralities in the co-occurrence sub-graph. We hypothesise that a combination of both measures -local weights representing the bookmark content topics and global co-occurrences taking into account collaborative popularities -is an appropriate strategy for tag recommendation.</p><p>Figure <ref type="figure" target="#fig_4">2</ref> shows the resultant co-occurrence graph associated to the tags retrieved from the example input bookmark. The tags with highest vertex in-degree seem to be good candidates to describe the contents of the bookmarked document. The goal of this stage was to establish global relations between the social tags that are potentially useful for describing the input bookmark. Exploiting these relations, we aimed to take into account tag popularity and tag co-occurrence aspects, and expected to identify which are the most informative tags to be recommended.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.5">Recommending social tags</head><p>In the fifth stage (label 5 in Figure <ref type="figure" target="#fig_0">1</ref>), we select and recommend a subset of the related tags from previous stages. The selection criterion we propose is based on three aspects: the tag frequency in bookmarks similar to the input bookmark (stage 3), the tag co-occurrence graph centrality (stage 4), and a personalisation strategy that prioritises those tags that are related to the input bookmark and belong to the set of tags already used by the user to whom the recommendations are directed.</p><p>For each tag ‫,ݐ‬ the first two aspects are combined as follows:</p><formula xml:id="formula_17">ܿ ሺ‫ݐ‬ሻ = ‫݁݁ݎ݃݁݀_݊݅‬ ሺ‫ݐ‬ሻ • ሺ‫ݒ‬ ሺ‫ݐ‬ሻሻ ଶ</formula><p>where ‫݁݁ݎ݃݁݀_݊݅‬ ሺ‫ݐ‬ሻ is the number of edges that have as destination the vertex of tag ‫ݐ‬ in the co-occurrence sub-graph built in stage 4 for the input bookmark ܾ . In order to penalise too generic tags we conduct a TF-IDF based reformulation of the centralities ܿ ሺ‫ݐ‬ሻ:</p><formula xml:id="formula_18">‫ݎ‬ ሺ‫ݐ‬ሻ = ܿ ሺ‫ݐ‬ሻ • ‫݈݃‬ ൬ ܰ #{݅: ‫ݐ‬ ߳ tagsሺܾ ሻ} ൰</formula><p>where ܰ is the total number of bookmarks in the repository. Finally, to take into account information about the user's tagging activity, we increase the ‫ݎ‬ ሺ‫ݐ‬ሻ values of those tags that have already been used by the user:</p><formula xml:id="formula_19">‫‬ ,௨ ሺ‫ݐ‬ሻ = ‫ݎ‬ ሺ‫ݐ‬ሻ • ሺ1 + ‫‬ ௨ ሺ‫ݐ‬ሻሻ</formula><p>where ‫‬ ௨ ሺ‫ݐ‬ሻ is the normalised preference of user ‫ݑ‬ for tag ‫:ݐ‬</p><formula xml:id="formula_20">‫‬ ௨ ሺ‫ݐ‬ሻ = ቐ ݂ ௨,௧ max ఢ ୲ୟୱሺ௨ሻ ݂ ௨, if ‫ݐ‬ ∈ ‫‪ሻ‬ݑ‪ሺ‬ݏ݃ܽݐ‬ 0 otherwise , ݂ ௨</formula><p>, being the number of times tag ‫ݐ‬ has been used by user ‫.ݑ‬ The tags with highest preference values ‫‬ ,௨ ሺ‫ݐ‬ሻ constitute the set of final recommendations. Both the TF-IDF and personalisation based mechanisms were evaluated isolated and in conjunction with the baseline approach ܿ ሺ‫ݐ‬ሻ improving its results.</p><p>Table <ref type="table" target="#tab_7">5</ref> shows the final sorted list of tags recommended for the example input bookmark: recommender, collaborative, filtering, semanticweb, personalization. It is important to note that these tags are not the same as the top tags obtained in Stage 3 (see Table <ref type="table" target="#tab_5">4</ref>). In that case, all those tags (recommender, recommendation, collaborative, filtering, collaborativefiltering) were biased to vocabulary about "recommender systems", and no diversity in the suggested tags was provided. In the fifth and last stage, we ranked the social tags extracted from the bookmarks similar to the input one. For that purpose, a combination of tag co-occurrence graph centrality, tag frequency, and tag-based personalisation metrics was performed. With an illustrative example, we showed that this strategy seems to offer more diversity in the recommendations than simply selecting the tags that more times were assigned to similar bookmarks. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Experiments</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1">Tasks</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2">Datasets</head><p>Table <ref type="table" target="#tab_10">6</ref> shows the statistics of the training and test datasets used in the experiments. Tag assignments (user-tag-resource) are abbreviated as tas. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3">Evaluation metrics</head><p>As evaluation metric, we use the average ‫-ܨ‬measure, computed over all the bookmarks in the test dataset as follows:</p><formula xml:id="formula_21">‫ݏ݃ܽݐ‪൫‬ܨ‬ ሺ‫,ݑ‬ ܾሻ൯ = 2 • ‫ݏ݃ܽݐ‪݊൫‬݅ݏ݅ܿ݁ݎ‬ ሺ‫,ݑ‬ ܾሻ൯ • ‫ݏ݃ܽݐ‪݈݈݁ܿܽ൫‬ݎ‬ ሺ‫,ݑ‬ ܾሻ൯</formula><p>‫ݏ݃ܽݐ‪݊൫‬݅ݏ݅ܿ݁ݎ‬ ሺ‫,ݑ‬ ܾሻ൯ + ‫ݏ݃ܽݐ‪݈݈݁ܿܽ൫‬ݎ‬ ሺ‫,ݑ‬ ܾሻ൯</p><p>where:</p><formula xml:id="formula_22">‫ݏ݃ܽݐ‪݈݈݁ܿܽ൫‬ݎ‬ ሺ‫,ݑ‬ ܾሻ൯ = ‫,ݑ‪ሺ‬ݏ݃ܽݐ‪ห‬‬ ܾሻ ∩ ‫ݏ݃ܽݐ‬ ሺ‫,ݑ‬ ܾሻห ‫,ݑ‪ሺ‬ݏ݃ܽݐ|‬ ܾሻ| ‫ݏ݃ܽݐ‪݊൫‬݅ݏ݅ܿ݁ݎ‬ ሺ‫,ݑ‬ ܾሻ൯ = ‫,ݑ‪ሺ‬ݏ݃ܽݐ‪ห‬‬ ܾሻ ∩ ‫ݏ݃ܽݐ‬ ሺ‫,ݑ‬ ܾሻห ‫ݏ݃ܽݐ‪ห‬‬ ሺ‫,ݑ‬ ܾሻห</formula><p>being ‫,ݑ‪ሺ‬ݏ݃ܽݐ‬ ܾሻ the set of tags assigned to bookmark ܾ by user ‫,ݑ‬ and ‫ݏ݃ܽݐ‬ ሺ‫,ݑ‬ ܾሻ the set of tags predicted by the tag recommender for bookmark ܾ and user ‫.ݑ‬ For each bookmark in the test dataset, we compute the ‫-ܨ‬measure by comparing the recommended tags against the tags the user originally assigned to the bookmark. The comparison is done ignoring case of tags and removing all characters which are neither letters nor numbers.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.4">Results</head><p>The tag recommendation approach presented in this work exploits training bookmark meta-information and tags, but does not analyse document contents, and does not make use of external knowledge bases, to enrich the set of suggested tags. Thus, all our recommended tags belong to the training collection, and our algorithm is only suitable for Task 2 of the ECML PKDD 2009 Discovery Challenge.</p><p>Table <ref type="table" target="#tab_11">7</ref> shows recall, precision and ‫-ܨ‬measure values for the test datasets provided in the tasks. In task 2, recommending 5 tags, we reach an average ‫-ܨ‬measure value of 0.3065. We obtain a precision of 42% if we only recommend one tag, and 25% when we recommend 5 tags. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Conclusions and future work</head><p>In this work, we have presented a social tag recommendation model for a collaborative bookmarking system. Our approach receives as input a bookmark (of a web page or a research publication), analyses and processes its textual metadata (document title, URL, abstract and descriptions), and suggests tags relevant to bookmarks whose metadata are similar to those of the input bookmark. Besides focusing on those tags that best fit the bookmark metadata, our strategy also takes into account global characteristics of the system folksonomy. More specifically, it makes use of the tag co-occurrence graph to compute vertex centralities of related tags. Assuming that tags with higher vertex centralities are more informative to describe the bookmark contents, our model weights the retrieved tags through their centrality values in a small co-occurrence sub-graph generated for the input bookmark. As additional features, the weighting mechanism also penalises tags that are too generic, and strengthens tags that have been previously used by the user to whom the tag recommendations are conducted.</p><p>Two are the main benefits of our approach: a low computational cost, and the capability of providing diversity in the recommended tag sets. On one hand, an index of keywords and tags for the available bookmarks, and the global tag co-occurrence graph, are the only information resources needed. On the other hand, the combination of exploiting content-based features, tag popularity and personalisation in the recommendation process allows suggesting tags that not only are relevant for the input bookmark, but also might belong to different domains.</p><p>A main drawback of our approach is its limitation to recommend tags that already exist in the system folksonomy. The suggestion of new terms, for example extracted from the bookmarked text contents or from external knowledge bases such as dictionaries or thesauri, is thus an open research line.</p><p>More investigation is needed to improve and evaluate the effectiveness of our tag recommender. In this context, the study of alternative graph vertex centrality measures (e.g. <ref type="bibr">[11]</ref>), and the exploitation of extra folksonomic information obtained from the user and item spaces (e.g., as done in <ref type="bibr">[6]</ref>), represent priority tasks to address in the future. The evaluation has to be also done comparing our approach with other state-of-the-art techniques.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Tag is a new form to index web resources, which help users to categorize and share the resources, and later search them. Also, the tags assigned by specified user reveal the user's interests, therefore, according to the tags user have already tagged, someone can find other users who have the similar interests, as well as similar interesting resources. Therefore, it is widely used in social network such as Bibsonmy, Del.icio.us, Last.fm , etc. A tag recommendation system can suggest someone a few tags to specified web resource, thus it can save the user time and effort when them mark up resources. Further, the recommended tags and existing tags can be used to predict the profile of the user and the interesting to the web resource, for example, to predict what they like and dislike. The research of tag recommendation is also very suggestive for other applications, such as online advertisement. In the field of online advertisement, we can predict what advertisement the browser might be interested in with the help of the surrounding text and his browsing history.</p><p>Recently, social tag recommendation has gained more attention in web research. It has been a hot issue for both industry and research area. For example, tag recommendation is one of the tasks in ECML RSDC's 08. Now, in ECML PKDD 09, tag recommendation has become the exclusive task. However, the performance of tag recommendation is not good enough to be widely used, more research work is needed and progress is essential for the practical use of tag recommendation in commercial system. In this paper, supervise ranking model is applied to tackle tag recommendation problem, and good result is achieved on test data.</p><p>The rest of paper is organized as follows: Section 2 lists the previous work on tag recommendation. Section 3 gives a description of supervised ranking model. Section 4 lists our experiment settings, experiment procedure and our analysis of the results on recovered 08's dataset. The model's performance on 09's dataset is presented in Section 5. Section 6 summarizes our work.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Previous Work</head><p>Much research work has been done for tag recommendation, most of which can be categorized into two types, one is rule-based, the other is classificationbased.</p><p>Rule based approach is used by many researchers. Lipczak <ref type="bibr">[1]</ref> proposed a three-step tag recommendation system in their paper : Basic tags are extracted from the resource title. In the next step, the set of potential recommendations is extended by related tags proposed by a lexicon based on co-occurrences of tags within resource's posts. Finally, tags are filtered by the user's personomy -a set of tags previously used by the user. Tatu, et al. <ref type="bibr">[2]</ref>used document and user models derived from the textual content associated with URLs and publications by social bookmarking tool users, the textual information includes information present in a URL's title, a user's description of a document, or a bibtex field associated with a scientific publication, they used natural language understanding approach for producing tag recommendations, such as extraction of concepts, extraction of conflated tags which group tags to semantically related groups. However, too much expert experience and manual work are needed in rule-based approaches, and its generalization is limited.</p><p>Classification-based approach is also used for the tag recommendation task. Katakis et al. <ref type="bibr">[3]</ref> tried to model the automated tag suggestion problem as a multilabel text classification task. Heymann et al. <ref type="bibr">[4]</ref> predicted tags based on page text, anchor text, surrounding hosts, and other tags applied to the URL. They found an entropy-based metric which captures the generality of a particular tag and informs an analysis of how well that tag can be predicted. They also found that tag-based association rules can produce very high-precision predictions as well as giving deeper understanding into the relationships between tags. Their results have implications for both the study of tagging systems as potential information retrieval tools, and for the design of such systems. However , the application of classification does not suggest a good solution to the tag prediction problem: first, the tag space is fixed , all the resource can be categorized to the existed tags only, also, the amount of tags number could be very large, the traditional classification model would be rather low efficient.</p><p>Collaborative filtering is a commonly used technical for user-oriented task. Many researchers tried collaborative filtering in tag recommendation. Gilad Mishne <ref type="bibr">[5]</ref> used collaborative approach to automated tag assignment for weblog posts. Robert Jaschke, et al <ref type="bibr">[6]</ref> evaluated and compared user-based collaborative filtering and graph-based recommender, the result shows that both of these two methods provide better results than non-personalized baseline method, especially the graph-based recommender outperforms existing methods considerably.</p><p>Adriana Budura et al. <ref type="bibr">[7]</ref> used neighborhood based tag recommendation, which make use of content similarity. Principle and simple score approach is used to select the candidate tags, however, in our paper, machine learning method is used, a ranking model is learned automatically, then the candidate tags are ranked and top-ranked tags are suggested as recommending tags.</p><p>3 Supervised Ranking Model for Tag Recommendation</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Problem Statement</head><p>The tag recommendation problem can be described as follows: For a given post P whose user is U and resource is R, a set of tags are suggested as tags for the post. Here we denote post as P, tag as T, resource as R, user as U.</p><p>A possible and most nature way to solve the tag recommendation problem is as follows: First, a set of candidate tags are selected for the post, and then tags which are most likely to be the tags for the post are selected as recommending tags. The commonly used approach to choose the tags is rule-based and classification-based methods, but both of them have defects: rule-based approach relies on expert experience and manual efforts to set up the rules and tuning the parameters; classification-based is restrict to the fix of tag space and is inefficient when it is treated as a multi-label problem. In this paper, tag recommendation is conveyed to a problem of ranking candidate tags. A ranking model is constructed to ensure tags that are most likely to be post's tags rank higher than tags that are not. Supervised learning model is used to construct the ranking model satisfying the restriction. Ranking-SVM model is the most frequently used supervised ranking model and is proofed to be a successful model, so it is used as our supervised ranking model in the experiments. All the candidate tags for one post are grouped as a ranking group and the top-ranked candidate tags are selected as recommendation tags.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Introduction to Ranking SVM</head><p>Here we briefly describe the Ranking Support Vector Machine(Ranking SVM) model for tag recommendation.</p><p>Assume that X ∈ ℜ m is the input feature space which represents feature of a candidate tag given a user and resource, and m denotes the feature number. Y = {0, 1} is the output rank space which is represented by the labels, and 1 represents the tag is labeled by user, and 0 is not. (x, y) ∈ X × Y denotes feature and label as the training instance.</p><p>Given a training set with tags T = {t 1 , t 2 , ..., t n }, for each tag t i there would be a {x, y} associated with it, the whole training set could be formulate as S = {x i , y i } N i=1 , where N represents the number of all tags. In Ranking SVM <ref type="bibr">[8]</ref>, ranking model f is a linear function represented by w, x , where w is the weight vector and •, • denotes the inner product. In RSVM we need to construct a new training set S ′ according to the original training set S = {x i , y i } N i=1 . For every y i = y j in S, construct (x i − x j , z ij ) and add it into S ′ , where z ij = +1 if y i ≻ y j , and otherwise −1. Here ≻ denotes the preference relationship, for example, y = 1 is preferred to y = 0. For denotation consistency, we denotes S ′ as {x</p><formula xml:id="formula_23">1 i − x 2 i , z i } D i=1 .</formula><p>The final model is formalized as the following Quadratic Programming problem:</p><formula xml:id="formula_24">min w,ξi 1 2C w 2 + D i=1 ξ i s.t. ξ i &gt; 0, z i w, x 1 i − x 2 i ≥ 1 − ξ i<label>(1)</label></formula><p>And ( <ref type="formula" target="#formula_1">1</ref>) could be solved using existing Quadratic Programming methods. Figure <ref type="figure" target="#fig_0">1</ref> is an example of ranking SVM model. The ranking SVM model convey the problem of ranking into binary classification problem: for each objects to be ranked, the model compare it with all other objects in the same ranking group. For n objects, the model compares the objects C 2 n times, and then outputs the ranking result.This is the advantage over classification model: in classification model, the existence of other candidate tags is not being considered, but in ranking model, the existence of other candidate tags is taken into consideration.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Ranking Process</head><p>For any post P ij in test dataset, we denote collection of all candidate tags for post P ij as CT {P ij } and CT k (k = 1, 2, ..., n) as the k-th candidate tag for the post P ij , CT {P ij } = {CT 1 , CT 2 , ..., CT n } . The ranking model ranks the candidate tags to {CT 1 ′ , CT 2 ′ , ..., CT n ′ } from top to bottom. Then top-k tags are selected as prediction of the tags of post P ij . Table <ref type="table" target="#tab_0">1</ref> shows the steps to rank the candidate tags. Also, the number of recommended tags affects the performance of the system. For example, if the actual number of tags for post whose content id=123456 is 3, a loss of precision is suffered when 4 tags are recommend to the user. So a proper number of tags to recommend should be found. The number used in our experiment is half the number of all candidate tags. If the number is bigger than 5, we cut them into 5, that means we recommend 5 tags at most.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4">Training Process</head><p>For all the post in the test dataset, candidate tags CT {P ij } for each post P ij are extracted. Then they are grouped by the post, and features are extracted for each of them in the post content. For those CT k ∈ T {P ij }, we label them '1', else label them with '0'. Then we use SVM-light tool to train a ranking-SVM model. When predicting the tags of the post in test dataset, the model learned on the training dataset is applied to rank the candidate tags, and top ranked tags are selected as recommending tags.</p><p>4 Experiments on 08's recovered dataset 4.1 Experiment settings 2008's dataset recovery In order to compare our experiments' performance with that of the 08's teams, we try to get the 08's dataset (both training and test data) and test our model's performance on the recovered dataset. Though the 08's test data can be downloaded from the web, we found that user IDs have been changed between the datasets. However, the content id field in 08's test data is consistent with 09's data, so we try to recover the 08's dataset on the 09's dataset using the content id field and date time field. The 08's real training data and test data are subset of 09's data, so it is possible to recover 08's data on 09's data. After observing 08's real test data, we found that all posts in 08's test data are between Mar. <ref type="bibr">31, 2008 and</ref><ref type="bibr">May. 15, 2009</ref>, so we use the posts during this period on 09's training data as recovered 08's test data and posts before Mar. 31, 2008 as our recovered 08's training data. There are still slight difference between our recovered data and the 08's real data. We assume that the difference won't affect our performance seriously, so the result is comparable with 08's results.</p><p>Some statistics have been made on our recovered 08's dataset. Table <ref type="table" target="#tab_1">2</ref> shows the statistics of posts on this recovered dataset. Table <ref type="table" target="#tab_3">3</ref> shows the statistics of posts according to the existence of their user and resource in the recovered training data. In following part in section 4, the training data refers to the recovered training data, the test data refers to the recovered test data.  Data preprocess Firstly, the terms are converted into lowercase. Then the stop words are removed, such as "a, the, is, an", these terms are not likely to be the tags of the post. Finally, the punctuations as ':', ',', etc are removed. Latex symbols such as '{' and '}' is also removed using regular expressions.</p><p>Table <ref type="table" target="#tab_7">5</ref> shows example results of data preprocess.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Post Division</head><p>It can be observed from data distribution that some users of posts exist in the training data (54%) and some do not exist in the training data (46%). Also  In the analysis above, we divide the posts in test dataset into two categories according to the existence of their users in the training data: existed user posts, non-existed user posts. Also, the posts in test dataset can be divided into two categories according to the existence of their resource in the training data: existed resource posts, non-existed resource posts.</p><p>The posts can be divided into four different categories according to their user status and resource status in the training data: existed user existed resource post, existed user non-existed resource post, non-existed user existed resource post, non-existed user non-existed resource post.</p><p>We denote symbols as shown in Table <ref type="table" target="#tab_10">6</ref> to simplify the language. Table <ref type="table" target="#tab_11">7</ref> and Table <ref type="table" target="#tab_19">8</ref> show statistics after our post division on our recovered 08's data. It can be observed from statistics that not every category of posts occupies the same ratio of the posts. In BOOKMARK, EUNR posts occupied about 82.80% of all BOOKMARK posts. In BIBTEX, NUNR posts occupied about 93.43% of all BIBTEX posts. In order to promote our model's performance on the test dataset, we should focus on those data which occupy high proportion of the posts, that is: EUNR posts of BOOKMARK and NUNR posts in BIBTEX.</p><p>After data division, the following steps are carried out for our tag recommendation task.</p><p>1. Extract candidate tags by different methods according to the category of post.</p><p>2. Rank the candidate tags, and select top ranked tags as recommendation tags.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">Candidate tags extraction</head><p>According to the statistics of the sources of the tags on the dataset, we can find that tags can be retrieved from three sources mainly: 1.The content information of the post, such as 'description' field in BOOKMARK and 'title' field in BIBTEX. 2. T {R j }: The tags being assigned to the same resource previously.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>3.T {U i }:</head><p>The tags assigned by the same user previously. Statistics of tags from different sources for BOOKMARK and BIBTEX posts are listed in Table <ref type="table" target="#tab_20">9</ref> and Table <ref type="table" target="#tab_21">10</ref>. The four different categories of test dataset have different characters, for example, we can explore the tags assigned by user previously and the tags assigned to the resource previously for EUER posts. But for NUNR posts, we lack this information. So we should explore different features for the four different categories of posts individually, in order that existed information can be used sufficiently. In the following part, while using the supervised ranking model, we train four models to handle these four categories of posts individually.</p><p>The candidate tags extraction strategies for different categories of posts: For EUER post and NUER post, CT {P ij } = { terms in post (P ij ) T {R j }}.</p><p>For EUNR post and NUNR post, CT {P ij } = { terms in post (P ij )}. We denote the candidate tags for post whose user id=i and resource is j as CT {P ij }. { terms in post (P ij )} denotes the remaining set of words after trimming and removing of the stop words in the text information of post P ij .</p><p>Notice should be paid here that we do not take T {U i } (the user's pervious tags) as candidate tags because we find the tags are too massive. When they are added, the precision of the system drops down and the F-1 value on the whole dataset also declines dramatically. However, in the ranking procedure, we will use T {U i } as one of the features in SVM model to rank the candidate tags.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4">SVM Features construction</head><p>While using SVM, we select features that discern high ranked tags and low ranked tags well and add the features according to our experience. For example, the term frequency in the post content: those words which have high term frequency within the post content tend to rank higher than those which have low term frequency. Also, whether the candidate words have been used as tags for other post in the training data is an excellent feature.</p><p>Table <ref type="table" target="#tab_22">11</ref> is a brief description of features of ranking SVM model for BOOK-MARK posts. The features for BIBTEX posts are almost the same except for the different data fields:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.5">Analysis of Model</head><p>Table <ref type="table" target="#tab_23">12</ref> and Table <ref type="table" target="#tab_24">13</ref> show the results of our supervised Ranking SVM model on the recovered 08's data.</p><p>Combing different types and category of data together, we can get the overall performance on the recovered 08's test data, as shown in Table <ref type="table" target="#tab_25">14</ref>.  The F1-value is 0.167, less than the F1-value 0.193 of the team ranked first in 08's competition.</p><p>It can be observed from the results that the performance of the model is poor on EUNR posts, which occupied most of the BOOKMARK posts. However, the model performs well on EUER posts. When comparing the two types of data, we find that the only difference is that the candidate tags of EUER posts are not only come from the post content but also from the tags of the same resource in the training data, however, the candidate tags for EUNR posts come from post content only. In order to overcome the weakness of lacking candidate tags, we relax restriction on the definition of the same resource. For those posts whose resources have not appeared in the training data, the role of the same post is substituted by the similar post. This method is based on the assumption that users tend to tag the similar posts with the same tags.</p><p>We try to use post content similarity to measure the similarity of posts. For those EUNR posts, which have no same resources in the training data, we add the tags of those posts whose content similarity with the current post content is above a certain threshold to the candidate tags set of the post.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.6">Post content similarity based KNN model</head><p>For EUNR post, the candidate tags come from text of the post content only, that is CT {P ij } = { terms in post (P ij )}. We attribute the poor performance of the model on such kind of data to the sparse of candidate tags. So we use content similarity to expand the candidate tags set. For any EUNR post P ij , we set a similarity threshold t, and find in the training dataset content P mn , whose sim(text(P ij ), text(p mn ) &gt; t). Then the tags of post P mn are added to the candidate tags ofP ij : CT {P ij } = { terms in post (P ij )} T {P mn }.</p><p>Post content P ij and P mn are mapped into vector space:</p><formula xml:id="formula_25">text(P ij ) = {W 1 , W 2 , ..., W n } , text(P mn ) = {W 1 ′ , W 2 ′ , .</formula><p>.., W n ′ },Then we use vector space model to calculate the similarity between two posts P ij and P mn .</p><formula xml:id="formula_26">sim(text(P ij ), text(P mn )) = text(P ij ) * text(P mn ) |text(P ij )| * |text(P mn )|<label>(2)</label></formula><p>W i means the weigh of word i in the content. The simplest way to define W i is as following:W i = 0,word i in post content, W i = 0,word i not in post content.</p><p>In our experiment, we define the W i as TF(Term Frequency) multiply IDF (Inverted document frequency) :W i = T F i * IDF i .We applied open source software Lucene to calculate the similarity of two content , the scoring function of Lucene is a derivation of vector space model formula using TF/IDF weighing schema.</p><p>The modification of threshold value T and the corresponding performance on EUNR content in BOOKMARK are shown in Figure <ref type="figure" target="#fig_4">2</ref>.</p><p>It can be observed that the value of recall, precision and F1 value reach highest when threshold T=0.5. So, in the further experiment settings, we set threshold value T to 0.5.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Fig. 2. KNN performance on various threshold t on BOOKMARK EUNR posts, k=5</head><p>However, we find that the application of content similarity based KNN model works for BOOKMARK posts but not for BIBTEX posts. After investigation, we attribute it to the uneven distribution of the dataset in training datasets and test datasets. In training datasets, the number of BOOKMARK posts is 184,655 and the number of BIBTEX posts is 20,647. But in test dataset, the number of BOOKMARK post is 20,647 and the number of BIBTEX post is 49,479, it is easy for 20,647 BOOKMARK posts to find similar posts in 184,655 BOOKMARK posts, but difficult for 42,545 BIBTEX posts in only 20,647 posts. So this method is especially useful for BOOKMARK posts but not for BIBTEX posts.</p><p>After applying content similarity based KNN model on BOOKMARK EUNR posts, the performance on overall test dataset is as listed in Table <ref type="table" target="#tab_26">15</ref>. The F1-value is 0.238, higher than the F1-value 0.193 of the team ranked first in 08s competition.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Experiment on 09's dataset</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1">Statistics of 09's dataset</head><p>Table <ref type="table" target="#tab_27">16</ref> and Table <ref type="table" target="#tab_28">17</ref> show the distribution of different categories of posts on 09's dataset after data division according to the existence of their user and resource in the training data. In our experiment settings on 09's test data, cleandump dataset is used as training dataset in Task 1, Post-core dataset is used as training dataset in Task 2. It can be observed from the statistics of the distribution of categories in 09's test data for Task 1 agrees with the recovered 08's dataset: EUER posts occupied most of the BOOKMARK post and NUNR post occupied large proportion of BIBTEX posts, so we can expect our model a good result on such data. The whole posts in 09's test dataset for Task2 can be classified to EUER posts. Since the good performance of our model on EUER posts, we can also expect a good result on task 2.</p><p>Eight different models are trained on 09's clean-dump training data and applied in 09's test data for Task 1. For Task 2, we apply the BOOKMARK EUER post model and the BIBTEX EUER post model trained on 09's postcore dataset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2">Experiment results on 09's test dataset</head><p>The performance on the whole 09's test data of both task 1 and task 2 is shown in Table <ref type="table" target="#tab_29">18</ref>. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Conclusion</head><p>In this paper, we briefly describe an approach utilizes supervised ranking model for tag recommendations. Our tag prediction contains three steps. First, posts are divided into four categories according to the existence of the user and the resource in the training data and then candidate tags are extracted for the different categories with different strategies. Second, features are decided according to categories. Then we rank the candidate tags, using the supervised ranking model, and pick the top tags as recommendation tags.</p><p>For the existed user non-existed resource post, we use post content similarity based KNN model to expand the candidate tags set. Performance of this experiment for the corresponding module is promoted after adding this model on 08's dataset. Our tag recommendation system is generated from the combination of these two models and applied to the 09's tags recommendation task 1 and task 2.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Tag Recommendations Based on Tracking Social</head><p>Bookmarking Systems</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>The development of collaborative society that we experience in recent years can be characterized by four principles: being open, peering, sharing and acting globally <ref type="bibr">[6]</ref>. These principles determine the way we exchange information and organize the knowledge. Very important part of this phenomenon is the popularity of social classification, indexing and tagging. Attaching labels to common resources (webpages, blogs, music, videos, photos) can on one hand shed a new light on information retrieval problems, on the other hand poses new challenges concerning uncontrolled explosion of folksonomy size and its usability. The goal of our research is to build a tag recommendation system that would influence user's selection of tags and as a result enable us to reuse folksonomy entries in more efficient way than we observe currently This paper describes our attempt to predict tags already chosen by BibSonomy users. This was the Task 1 in ECML PKDD 2009 Challenge. However, we believe that our system is better suited for the third Task, in which the Teams have an opportunity to deliver recommendations online.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>2</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Related work</head><p>The growing interest of research community in the field of social bookmarking was fueled by last year's ECML challenge, during which the evaluation measures were standardized and benchmark data sets prepared. Thirteen solutions were submitted to the tag spam detection task and only five to tag recommendation task. We were inspired by the best teams in the Challenge, which relied on several external resources <ref type="bibr">[7]</ref> and used only data available in title/description fields <ref type="bibr">[5]</ref>. The team from the Aristotle University of Thessaloniki <ref type="bibr">[3]</ref> reformulated the task as a multilabel classification problem.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Examined datasets</head><p>We used cleaned dump dataset which consisted of three tables: bibtex (158 924 records), bookmark (263 004 records) and tas (1 401 104 records). The dump contained all public bookmarks and publication posts of BibSonomy until (but not including) 2009-01-01. Posts from the user dblp (a mirror of the DBLP Computer Science Bibliography) as well as all posts from users which have been flagged as spammers have been excluded. Furthermore, the tags were cleaned. Java method was used to remove all characters which were neither numbers nor letters and removed those tags, which were empty after cleansing or matched one of the tags imported, public, systemimported, nn, systemunfiled. The tas table (Tag Assignments) was a fact table with information about who attached which tag to which resource/content. The bookmark table consisted of following columns (content_id, url_hash, url, description, extended description and date). The bibtex table was described by following dimensions (content_id, journal, volume, chapter, edition, month, day, booktitle, howPublished, institution, organization, publisher, address, school, series, bibteXKey, url, type, description, annote, note, pages, key, number, crossref, misc, bibtexAbstract, simhash0, simhash1, simhash2, entrytype, title, author, edition, year).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Our approach</head><p>In this section we describe three main parts of our system. Firstly we focus on a selection of RSS feeds and the problems we encountered while downloading the posts. In the second part we define the vector space in which the posts were stored as well as main characteristics of deployed database. Finally we present the details of the tag recommendation algorithm. The algorithm is divided into four steps: searching of matching resources based on URL address, retrieval of the most similar cluster, selection of the post with highest overlap score and ranking of suggested tags.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>4.1</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>RSS Feeds selection</head><p>Our strategy was to optimize a set of keywords that we were going to track in popular bookmarking systems as well as in a variety of domain portals. We analyzed distribution of most common tags in BibSonomy and Delicious and decided that tracking only the most recent posts would be biased (Table <ref type="table" target="#tab_0">1</ref>). We decided to enrich the most recent posts with a set of 100 most popular tags (out of 93 757 unique tags) in BibSonomy training data. We had to face different problems in case of bookmarking systems and domain portals. We used Google Reader to search for top 10 domain portals and their RSS URLs for each chosen keyword. Because some feeds appeared in different searching results we end up with 734 feeds.</p><p>An example of feeds recommended by Google Reader for a keyword "linux" is presented in Table <ref type="table" target="#tab_1">2</ref>. Even though numerous feeds use the most recent RSS or Atom standard and we could easily parse the content of XML files, it is uncommon to fill in the category field by feed editors. We can see in the Table <ref type="table" target="#tab_1">2</ref>, that out of 10 sources: one did not contain proper URL, four did not deliver information about category, one marked each feed entry with the same category. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>RSS Feed</head><p>Categories of updated entries On the other hand the problem with typical bookmarking systems is the fact that when we subscribe most recent posts for a given keyword we get only tags of a particular user who bookmarked the resource. As a consequence we need to crawl a service in order to find out about most typical tags for a given resource. The problem of connection limits arises when we want to crawl every out of 100 entries downloaded for a given keyword. Because of this, we decided to verify if we can cluster tags based on their cooccurence score. Table <ref type="table" target="#tab_3">3</ref> contains 20 pairs of tags with highest symmetric Jaccard cooccurance coefficient calculated as a division of number of posts with both tags by a number of all posts with any of the tags. We can see that "ccp" and "jrr" always appear together. Also "genetic", "algorithms" and "programming" create a cloud of tags. Four tags "emulationgames", "emulationvideogames", "aaaemulationgames", "classicemulatedremakeretrogames" create another cloud. However, the Jaccard coefficient drops very fast below 20% level and therefore we decided not to abandon the idea of tag clustering.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Linux</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>4.2</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Data storage</head><p>In order to recommend tags online we needed a fast engine that does not need to be taught every time we get need posts from a scratch. The Beatca system (developed in our Institute <ref type="bibr">[1,</ref><ref type="bibr">4]</ref>) is an example of such engine. It performs online incremental hierarchical clustering of documents and proved very effective in the field of intelligent Information Retrieval. Soft classification of documents and construction of conceptual closeness graph is based on large-scale Bayesian networks. Optimal document map search and document clustering is based on SOM (self-organizing maps), AIS (artificial immune systems), and GNG (growing neural gas). Each post is defined as a point in a multidimensional space in which coordinates represent frequency of a token appearing in a post's title or description. Because some tokens are very common and others are present in only few posts we selected only the most informative tokens as coordinates in our vector space. The dictionary optimization was based on a entropy-like quality measure Q(t i ) of a token t i :</p><formula xml:id="formula_27">˝{ˮ { ˚ ˚Ñ . ˚ ˚ Ñ ˚ ˚ (# ˚<label>(1)</label></formula><p>where N ij is the number of occurrences of term t i in document d j , N j is the number of documents that contains term t i and N is the total number of documents. We removed tokens with Q(t i ) measure below 0.01 or above 0.95. We implemented term frequency inverse document frequency weighting scheme. According to the scheme we divided term frequency in a single document by the number of documents in which the term appears.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">Tag recommendations</head><p>Our tag recommendation consisted of four steps. If we had a positive result in the first step then we went directly to the final fourth step.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Step One</head><p>In the first step we checked if a post is present in the BibSonomy training set or an URL of the post is among downloaded RSS entries. If the answer was true then we selected all tags attached to these resources and moved to the Step Four.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Step Two</head><p>In the second step we retrieved a group (cluster) of documents that was the most similar to the post's description or title field. The similarity was measured as a cosine of an angel between vectors x={x 1 ,…,x n } and y={y 1 ,…,y n } representing the resources in our database and the post (Eq. 2). For example, one of the posts had following title: "Attribute Grammar Based Programming and its Environment". The query consisting of the first five informative tokens from the above title returned a cluster of four documents:</p><p>1 </p><formula xml:id="formula_28">{ { ˲ ˩ ˳ ˩ J ˩ ŵ ˲ ˩ Ŷ J ˩ ŵ ˳ ˩ Ŷ J ˩ ŵ<label>(2)</label></formula><p>A cluster of all the retrieved posts was transferred to the next step.</p><p>Step Three For all the posts retrieved in the second step we calculated normalized overlap score and chosen the post with the highest score. The overlap was defined as a maximum length of n-gram appearing in both posts. In order to compute the score we used all the words from title/description fields (not only the most informative tokens). The overlap score was divided by the length of title/description field of the candidate posts. For example, normalized overlap score between "Attribute Grammar Based Programming and its Environment" and "Attribute grammar based language extension for Java" equals to 3/7=0.42. The post with highest score was transferred to the final fourth step if the value of a score was greater than 0.6 threshold.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Step four</head><p>In the last step we ordered the tags of selected post according to their count in BibSonomy training set. Top five tags were selected as predictions in the Challenge.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Evaluation</head><p>The F1-Measure common in Information Retrieval was used to evaluate the recommendations. The precision and recall were first computed for each post in the test data by comparing the recommended tags against the tags the user has originally assigned to this post <ref type="bibr">[2]</ref>. Then the average precision and recall over all posts in the test data was used to calculate the F1-Measure as f1 = (2 * precision * recall) / (precision + recall). The number of tags one can recommend was not restricted. However, the organizers regarded the first five tags only. We computed both precision and recall measures for various levels of a threshold parameter from step three in our recommendation algorithm (Fig. <ref type="figure" target="#fig_0">1</ref>). According to these simulations optimum level of the threshold is approximately 0.6 and yields F1-measure between 3% and 4%. During the challenge we obtained overall F1-measure of 4,6%, which was slightly better than in our simulations, but incomparable to the results of the best teams.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Conclusions</head><p>We must admit that the way we approached the problem needs substantial computing power and disc space. Unfortunately the quality of our tag recommendations was below an average and probably this direction of research in the field of tag recommending systems is not a promising one. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Collaborative tagging has emerged as a popular method for organizing and sharing online content with user-defined keywords. Delicious<ref type="foot" target="#foot_16">1</ref> , Flickr<ref type="foot" target="#foot_17">2</ref> and Last.fm<ref type="foot" target="#foot_18">3</ref> are among the most popular destinations on the Web allowing users to annotate bookmarks, digital photographs and music. Other less popular tagging applications serve niche communities enabling users to tag blogs, business documents or scholarly articles. At the heart of collaborative tagging is the post; a user describes a resource with a set of tags. A collection of posts results in a complex network of interrelated users, resources and tags commonly referred to as a folksonomy <ref type="bibr">[10]</ref>.</p><p>The rich tapestry of a folksonomy presents an enticing target for data mining techniques such as recommenders. Recommenders reduce a burdensome number of items to a manageable size correlated to the user's interests. Recommendation in folksonomies can include resources, tags or even other users. In this work we focus on tag recommendation, the suggestion of tags during the annotation process.</p><p>Tag recommendation reduces the cognitive effort from generation to recognition. Users are therefore encouraged to tag more frequently, apply more tags to a resource, reuse common tags and perhaps use tags the user had not previously considered. User error is reduced by eliminating capitalization inconsistencies, punctuation errors, misspellings and other discrepancies. The final result is a cleaner denser dataset that is useful in its own right or for further data mining techniques.</p><p>Despite the richness folksonomies offer, they present unique challenges for tag recommenders. Traditional recommendation strategies, often developed to work with two dimensional data, must be adapted to work with the three dimensional nature of folksonomies. Otherwise they risk disregarding potentially useful information. To date the most successful tag recommenders are graph-based models, which exploit the links between users, resources and tags. However, this approach is computationally intense and ill suited for large scale implementation.</p><p>In this work we propose a composite tag recommender incorporating several distinct recommendation strategies. These recommenders are combined to generate a new hybrid. As such no single recommender is required to fully exploit the data structure of the folksonomy. Instead the recommenders may specialize in a single channel. The aggregation of these recommenders, none of which performs well on its own, produce a synergy allowing the composite recommender to outperform its constituent parts.</p><p>Our hybrid includes popularity models and item-based collaborative filtering techniques. Popularity based approaches include information garnered from the crowd with little computational cost. Item-based collaborative filtering focuses more closely on the user's profile incorporating a degree of personalization.</p><p>We provide a through evaluation of the composite recommender and its constituent parts. Our experiments reveal that the composite model produces results far superior to the capabilities of their individual components. We further include a comparison with the highly effective but computationally inefficient graph-based approach. We show that a low cost alternative can be constructed from less time consuming recommenders and perform nearly as well as the state or the art graph based approaches.</p><p>The rest of the paper is organized as follows. In Section 2 we discuss related work. In Section 3 we offer a model of folksonomies and describe tag recommendation. We further describe four recommendation algorithms. Informational channels in folksonomies are discussed in Section 4. We design a hybrid recommender in Section 5. Our experimental evaluation is presented in Section 6 including a discussion of the dataset, methodology and results. Finally we end the paper with a discussion of our conclusions and directions for future work in Section 7.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Related Work</head><p>As collaborative tagging applications have gained in popularity researchers have begun to explore and characterize the tagging phenomenon. In <ref type="bibr">[9]</ref> and <ref type="bibr">[4]</ref> the authors studied the information dynamics of Delicious, one of the most popular folksonomies. The authors discussed how tags have been used by individual users over time and how tags for an individual resource stabilize over time. They also explored two semantic difficulties: tag redundancy, when multiple tags have the same meaning, and tag ambiguity, when a single tag has multiple meanings. In <ref type="bibr">[9]</ref> the authors provide an overview of the phenomenon and explore reasons why both folksonomies and ontologies will have a place in the future of information access.</p><p>There have been several recent research investigations into recommendation within folksonomies. Unlike traditional recommender systems which have a twodimensional relation between users and items, tagging systems have a three dimensional relation between users, tags and resources. Recommender systems can be used to recommend each of the dimensions based on one or two of the other dimensions. In <ref type="bibr">[17]</ref> the authors apply user-based and item-based collaborative filtering to recommend resources in a tagging system and uses tags as an extension to the user-item matrices. Tags are used as context information to recommend resources in <ref type="bibr">[13]</ref> and <ref type="bibr">[12]</ref>.</p><p>Other researchers have studied tag recommendation in folksonomies. In <ref type="bibr">[7]</ref> user-based collaborative filtering is compared to a graph-based recommender based on the Pagerank algorithm for tag recommendation. The authors in <ref type="bibr">[5]</ref> use association rules to recommend tags and introduce an entropy-based metric to define how predictable a tag is. In <ref type="bibr">[8]</ref> the title of a resource, the posts of a resource and the user's vocabulary are used to recommend tags.</p><p>General criteria for a good tagging system including high coverage of multiple channels, high popularity and least-effort are presented in <ref type="bibr">[18]</ref>. They categorize tags as content-based tags, context-based tags, attribute tags, subjective tags, and organizational tags and use a probabilistic method to recommend tags. In <ref type="bibr">[2]</ref> the authors propose a classification algorithm for tag recommendation. The authors in <ref type="bibr">[15]</ref> use a co-occurrence-based technique to recommend tags for photos in Flickr. The assumption is that the user has already assigned a set of tags to a photo and the recommender uses those tags to recommend more tags. Semantic tag recommendation systems in the context of a semantic desktop are explored in <ref type="bibr">[1]</ref>. Clustering to make real-time tag recommendation is developed in <ref type="bibr">[16]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Tag Recommendation</head><p>Here we first provide a model of folksonomies, then review several common recommendation techniques which we employ in our evaluation. A folksonomy can be described as a four-tuple D = U, R, T, A , where, U is a set of users; R is a set of resources; T is a set of tags; and A is a set of annotations, represented as user-tag-resource triples: A ⊆ { u, r, t : u ∈ U, r ∈ R, t ∈ T }. A folksonomy can, therefore, be viewed as a tripartite hyper-graph <ref type="bibr">[11]</ref> with users, tags, and resources represented as nodes and the annotations represented as hyper-edges connecting a user, a tag and a resource.</p><p>Aggregate projections of the data can be constructed, reducing the dimensionality but sacrificing information <ref type="bibr">[14]</ref>. The relation between resources and tags, RT , can be formulated such that each entry, RT (r, t), is the weight associated with the resource, r, and the tag, t. This weight may be binary, merely showing that one or more users have applied that tag to the resource. In this work we assume RT (r, t) to be the number of users that have applied t to the r: RT tf (r, t) = |{a = u, r, t ∈ A : u ∈ U }|. Analogous two-dimensional projections can be constructed for UT in which the weights correspond to users and tags, and UR in which the weights correspond to users and resources.</p><p>Many authors have attempted to exploit the data model for recommendation in folksonomies. In traditional recommendation algorithms the input is often a user, u, and the output is a set of items, I. Tag recommendation differs in that the input is both a user and a resource. The output remains a set of items, in this case a set of recommended tags, T r . Given a user-resource pair, the recommendation set is constructed by calculating a weight for each tag, w(u, r, t), and recommending the top n tags.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Popularity Based Approaches</head><p>We consider two popularity based models which rely on the frequency a tag is used. PopRes ignores the user and relies on the popularity of a tag within the context of a particular resource. We define the resource based popularity measure as:</p><formula xml:id="formula_29">w(u, r, t) = |{a = u, r, t ∈ A : u ∈ U }| |{a = u, r, t ∈ A : u ∈ U, t ∈ T }|<label>(1)</label></formula><p>PopUser, on the other hand, ignores the resource and focuses on the frequency of a tag within the user profile. We define the user based popularity measure as:</p><formula xml:id="formula_30">w(u, r, t) = |{a = u, r, t ∈ A : r ∈ R}| |{a = u, r, t ∈ A : r ∈ R, t ∈ T }|<label>(2)</label></formula><p>Popularity based recommenders require little online computation. Models are built offline and can be incrementally updated. However both these models focus on a single channel of the folksonomy and may not incorporate otherwise relevant information into the recommendation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Item-Based Collaborative Filtering</head><p>KNN RT models resources as a vector over the tag space. As before the weights of the vectors may be calculated through a variety of means. Given a resource and a tag, we define the weight as the entry of the two dimensional projection, RT (r, t), the number of times r has been tagged with t.</p><p>When a user selects a resource to annotate, the similarity between it and every resource in the user profile is calculated. A neighborhood of the k most similar resources, S, is thus constructed. We then define the item-based collaborative filtering measure as:</p><formula xml:id="formula_31">w(u, r, t) = S s sim(s, r) * d(u, s, t) k<label>(3)</label></formula><p>where d(u, s, t) is 1 if the user has applied t to s and 0 otherwise. Like popUser, this recommender focuses strongly on the user's tagging practice. However this recommender includes an additional informational channel, identifying resources in the user profile that are similar to the query resource. This technique therefore includes resource-to-resource information.</p><p>If the system waits to compute the similarity between resources until query time, this recommender will also scale well to larger datasets so long as user profiles remain small. Alternatively similarities between resources can be computed offline. Consequently the computation at query time is dramatically reduced and the algorithm becomes viable for large collaborative tagging implementations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Folkrank</head><p>Folkrank was proposed in <ref type="bibr">[6]</ref>. It computes a Pagerank vector from the tripartite graph of the folksonomy. This graph is generated by regarding U ∪ R ∪ T as the set of vertices. Edges are defined by the three two-dimensional projections of the hypergraph, RT , U R and U T .</p><p>If we regard the adjacency matrix of this graph, W , (normalized to be column-stochastic), a damping factor, d, and a preference vector, p, then we iteratively compute the Pagerank vector, w, in the usual manner: w = dAw+(1−d)p.</p><p>However due to the symmetry inherent in the graph, this basic Pagerank may focus too heavily on the most popular elements. The Folkrank vector is taken as a difference between two computations of Pagerank: one with and one without a preference vector. Tag recommendations are generated by biasing the preference vector towards the query user and resource <ref type="bibr">[7]</ref>. These elements are given a substantial weight while all other elements have uniformly small weights.</p><p>We include this method as a benchmark as it has demonstrated to be an effective method of generating tag recommendations. However, it imposes steep computational costs. The channel between resources and tags reveals a highly descriptive model of the resources. The accumulation of many users' opinions (often numbered in the thousands or millions) results in a richness which taxonomies are unable to approximate. Conversely the tags themselves are characterized by the resources to which they have been assigned.</p><p>As users annotate resource with tags they define their interests in as much as they describe a resource. The user-tag channel therefore reveals the users interests and provides opportunities for data mining algorithms to offer a high degree of personalization. Likewise a user may be defined by the resources which he has annotated as in the user-resource channel.</p><p>These primary channels can be used to produce secondary informational channels. The user-user channel can be constructed by modeling users as a vector of tags or as a vector of resources and applying a similarity measure such as cosine similarity. Many variations exist. However the result reveals a network of users that can be explored directly or incorporated into further data mining approaches. The resource-resource and tag-tag channels provide similar utility, presenting navigational opportunities for users to explore similar resources or tags.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">A Multi-Channeled Tag Recommender</head><p>The most successful tag recommenders to date have included multiple informational channels. Folkrank explicitly includes the user-resource, user-tag and resource-tag channels in the graph model. Moreover since the algorithm calculates the Pagerank vector of the graph it implicitly includes the secondary channels of the folksonomy. The success Folkrank has achieved is due to its ability to incorporate multiple informational channels into a single tag recommender.</p><p>However the success it has achieved is blunted by the computational effort required to produce a recommendation; a new Pagerank vector is computed for each query.</p><p>Here we construct a hybrid recommender. The constituent parts by themselves perform poorly when compared to Folkrank. However, when aggregated into a single recommender they achieve a synergy which exploits several channels of the folksonomy while retaining their modest computational needs.</p><p>Our model includes PopRes, PopUser and KNN RT. We employ a weighted approach to combine the recommenders. First in order to ensure that weight assignments are on the same scale for each recommendation approach, we normalize the weights given to the tags by w(u, r, t) to 1 producing w (u, r, t). We then combine the weights in a linear combination: w(u, r, t) = αw P opRes (u, r, t) + βw P opU ser (u, r, t) + γw KN N RT (u, r, t)</p><p>such that weights α + β + γ = 1 and all values are positive. If α is set near 1 then hybrid would rely mostly on PopRes.</p><p>Tags promoted by PopRes will have a strong relevance to the resource, while tags promoted by PopUser will include tags in the user's profile. PopRes alone will ignore personal tags that the user often users. PopUser, on the other hand, will ignore tags related to the context of the query resource. Together these recommenders can include both aspects in the recommendation set. Moreover by including KNN RT tags which the user has applied to resources similar to the query resource are promoted.</p><p>PopRes explicitly includes the resource-tag information. PopUser, on the other hand, includes user-tag information. Both these models are based on popularity and are single-minded in their approach ignoring all data except the informational channel to which they are employed. We use KNN RT to introduce more subtlety into the hybrid. It focuses heavily on the user-tag channel, but gives more weight to tags that have been applied to similar resources. Hence it also includes resource-tag information. Moreover by focusing exclusively on resources in the user profile it includes the user-resource channel. Finally, KNN RT includes resource-resource information when it calculates the neighborhood of similar resources. This hybrid does not include user-user information or tag-tag information. Additional recommenders could be included to cover these informational channels. However, we have built this hybrid with the goals of speed and simplicity. The two popularity based approaches are among the fastest and simplest recommendation algorithms. The item-based collaborative filtering recommender is used to tie together these approaches incorporating similarities among resources into the model while retaining its speed. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Experimental Evaluation</head><p>In this section we describe the dataset used for experimentation. We then describe our experimental methodology and metrics. Finally we discuss the results of our experiments.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.1">Data Set</head><p>The dataset was provided by Bibsonomy<ref type="foot" target="#foot_20">4</ref> for the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML-PKDD) 2009 Challenge. BibSonomy was originally launched as a collaborative tagging application allowing users to organize and share scholarly references. It has since expanded its scope allowing users to annotate URLs.</p><p>The data includes all public bookmarks and publication posts of BibSonomy until 2009-01-01. The data was cleaned by removing all characters which are neither numbers nor letters from tags. Additionally the system tags imported, public, systemimported, nn and systemunfiled where removed.</p><p>Task 1 for the 2009 Challenge utilizes the complete dataset. Task 2 however focuses on the post-core at level 2 geared toward graph based approaches. For the post-core all users, tags, and resources which appear in only one post were removed. This process was repeated until convergence and produced a core in which each user, tag, and resource occurs in at least two posts. Reducing a dataset to its core was first proposed in <ref type="bibr">[3]</ref>. In <ref type="bibr">[6]</ref> it was adapted for folksonomies. The experiments for this work rely on post-core at level 2.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.2">Experimental Methodologies</head><p>We employ the leave one post out methodology as described in <ref type="bibr">[7]</ref>. One post from each user was placed in the testing set consisting of a user, u, a resource, r, and all the tags the user has applied to that resource. These tags, T h , are analogous to the holdout set commonly used in Information Retrieval evaluation. The remaining posts are used to generate the recommendation models.</p><p>The tag recommendation algorithms accepts the user-resource pair and returns an ordered set of recommended tags, T r . From the holdout set and recommendation set utility metrics were calculated. For each metric the average value was calculated across all test cases.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.3">Experimental Metrics</head><p>Recall is a common metric of recommendation algorithms that measures coverage. It measures the percentage of items in the holdout set, T h , that appear in the recommendation set T r . It is defined as:</p><formula xml:id="formula_33">r = (|T h ∩ T r |)/|T h |<label>(5)</label></formula><p>Precision is another common metric that measures specificity and is defined as:</p><formula xml:id="formula_34">p = (|T h ∩ T r |)/|T r |<label>(6)</label></formula><p>In order to conform to the evaluation methods of the ECML-PKDD 2009 Challenge, we use the F1-Measure common in Information Retrieval to evaluate the recommendations. We compute for each post the recall and precision for a recommendation set of five tags. Then we average precision and recall over all posts in the test data and use the resulting precision and recall to compute the F1-Measure as:</p><formula xml:id="formula_35">f 1 = (2 * p * r)/(p + r)<label>(7)</label></formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.4">Experimental Results</head><p>Our approach required that several variables be tuned. For KNN RT, after extensive experimentation of k in increments of 1 we set k equal to 15. We observed that as k increased from 0 to 15 recall and precision both increased rapidly until it suffers from diminishing returns. We evaluated the weights α, β and γ in .05 increments attempting every possible combination. Best results were found when α = 0.35, β = 0.15 andγ = 0.50. As such KNN RT accounts for 50% of the model, PopRes acounts for 35% and PopUser acounts for 15%.</p><p>KNN RT identifies resources in the user profile most similar to the query resource and promotes the tags applied to these resources. This approach is most effective when the user has generated a large user profile. Since users often employ tags as an organizational tool they often reuse tags. Hence the success of KNN RT stems from its ability to identify which previously used tags are most appropiate given the context of the query resource.</p><p>PopRes, on the other hand, ignores the user profile and concentrates on the popularity of a tag given the query resource. When the tags provided by KNN RT are insufficient, perhaps because the user has yet to build a deep user profile or is tagging a resource dissimilar to items in the profile, PopRes is able to provide relevant suggestions.</p><p>Finally PopUser promotes tags in the user profile regardless of the similarity to the query resource. It may promote idiosyncratic, subjective or organizational tags that do not necessarily relate to the context of the query resource but are often applied by the user.</p><p>Our evaluation of the composite recommenders in Figures <ref type="figure" target="#fig_52">2 and 3</ref> reveals that PopRes, PopUser and KNN RT achieve only modest success when used alone. However when combined together as a hybrid recommender the three are able to cover multiple informational channels and produce a synergy allowing the hybrid to produce superior results.</p><p>Not only is the hybrid recommender able to outperform the baseline recommenders it is also able to outperform Folkrank, a highly effective tag recommender. Moreover the hybrid retains the computationally efficiency of its parts making it suitable for deployment in large real work collaborative filtering applications.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7">Conclusions and Future Work</head><p>In this paper we have introduced the idea of informational channels in folksonomies and have proposed a fast yet effective tag recommender composed of three separate algorithms. The constituent recommenders were chosen for their speed and simplicity as well as their ability to cover complimentary informational channels. We have demonstrated that these recommenders while performing poorly alone, create a synergy when combined in a linear combination. The hybrid recommender is able to surpass the effective graph based approaches while retaining the efficiency of its parts. Future work will include an examination of alternative hybrid recommenders and present work on other datasets.</p><p>A modified and fast Perceptron learning rule and its use for Tag Recommendations in Social Bookmarking Systems 1 Introduction</p><p>Text categorization is the process of making binary decisions about related or non-related documents to a given set of predefined thematic topics or categories. This task is an important component in many information management organizations. In our participation on the ECML/PKDD challenge 2009, we treat Task 2 as a standard text classification problem and try to solve it using a machine learning, supervised, automatic classification method. The rest of the paper is organized as follows. Section 2 provides a description of the algorithm that we used. Section 3 briefly present the tasks of this year's Challenge. Section 4 presents the experimental setup, data processing and results and finally in section 5 we conclude on the results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">The Learning Algorithm</head><p>The algorithm that we used is an evolution of the algorithm that appeared in <ref type="bibr">[1]</ref> as a text classification algorithm and then a revised version in <ref type="bibr">[2]</ref> won last year's ECML PKDD Discovery Challenge on Spam Detection. The proposed algorithm is a binary linear classifier and it combines a centroid with a batch perceptron classifier and a modified perceptron learning rule that does not need any parameter estimation. Details on this modified algorithm, its experimental evaluation, theoretical investigation etc, have already submitted and are under review for publication at the time this paper was written. In the following paragraphs we will briefly describe this method that we used for solving the problem of ECML PKDD 2009 Discovery Challenge, Task 2.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Linear Classifiers</head><p>Linear Classifiers is a family of classifiers whose trained model is a linear combination of features. In another perspective linear classifiers train a model which is a hyperplane in a high dimensional feature space. In this space each instance, either of the train set or an unseen, is a point. The goal of a linear classifier is then to find such a hyperplane that splits the space into two subspaces, where one contains all the points of the positive class and the other contains all the points of the negative class.</p><p>Assuming that feature space is of n dimensions, each instance x i will be represented by an n dimensions vector</p><formula xml:id="formula_36">− → x i = (w i1 , w i2 , • • • , w in ) (1)</formula><p>where w ik is a real value of the kth feature for instance x i . Apart of each vector representation − → x i , each instance x i may bears information about being a member of a class or not. For example a document is known to be spam or an image is known that shows a benign tumor. This information can be coded using a variable y i for each instance x i which takes values as:</p><formula xml:id="formula_37">y i = 1 if x i ∈ C + −1 if x i ∈ C −<label>(2)</label></formula><p>That is y i = 1 when x i is member of the positive class C + and y i = −1 when it is member of the negative class C − . So each instance x i is represented by a tuple ( − → x i , y i ). A training set T r would be</p><formula xml:id="formula_38">T r = {( − → x 1 , y 1 ) , ( − → x 2 , y 2 ) , • • • , ( − → x m , y m )}<label>(3)</label></formula><p>A linear classifier then is defined by a model − → W , b where − → W is a vector in the same n-dimensional space and b is a scalar bias (threshold) value. This model defines a hyperplane h h :</p><formula xml:id="formula_39">− → W • x + b = 0 (4)</formula><p>This is the equation of a hyperplane h in the n-dimensional space. This hyperplane is of n − 1 dimensions. − → W is a linear combination of n features (dimensions). Hyperplane h splits space into two subspaces, the one where for every vector − → x i : </p><formula xml:id="formula_40">− → W • − → x i + b &gt;</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Perceptron</head><p>Perceptron is a flavor of Linear Classifiers. It starts with an initial model and iteratively refines this model using the classifications errors during training. It is the elementary particle of neural networks and it has been investigated and studied since the 1950s <ref type="bibr">[3]</ref>. It has been shown that when trained on a linearly separable set of instances, it converges (it finds a separating hyperplane) in a finite number of steps <ref type="bibr">[4]</ref> (which depends on the geometric characteristics of the instances on their feature space).</p><p>The Perceptron is a Linear Binary Classifier that maps its input − → x (a realvalued vector) to an output value f ( − → x ) (a single binary value) as:</p><formula xml:id="formula_41">f ( − → x ) = 1 if − → W • − → x + b &gt; 0 −1 else<label>(5)</label></formula><p>where − → W is a vector of real-valued weights and − → W • − → x is the dot product (which computes a weighted sum). b is the bias, a constant term that does not depend on any input value. The value of f ( − → x ) (1 or −1) is used to classify instance x as either a positive or a negative instance, in the case of a binary classification problem. The bias b can be thought of as offsetting the activation function, or giving the output neuron a "base" level of activity. If b is negative, then the weighted combination of inputs must produce a positive value greater than −b in order to push the classifier neuron over the 0 threshold. Spatially, the bias alters the position (though not the orientation) of the decision boundary (separating hyperplane h).</p><p>We can always assume for convenience that the bias term b is zero. This is not a restriction since an extra dimension n + 1 can be added to all the input vectors − → x i with − → x i (n + 1) = 1, in which case − → W (n + 1) replaces the bias term. Learning is modeled as the weight vector − → W being updated for multiple iterations over all training instances. Let</p><formula xml:id="formula_42">T r = {( − → x 1 , y 1 ) , ( − → x 2 , y 2 ) , • • • , ( − → x m , y m )}</formula><p>denote a training set of m training examples (instances). At each iteration k the weight vector is updated as follows. For each ( − → x i , y i ) pair in T r</p><formula xml:id="formula_43">− → W (k) = − → W (k−1) + α (k−1) 2 y i − f (k−1) ( − → x i ) − → x i (<label>6</label></formula><formula xml:id="formula_44">)</formula><p>where α is a constant real value in the range 0 &lt; α ≤ 1 and is called the learning rate. Note that equation 6 means that a change in the weight vector − → W will only take place for a given training example ( − → x i , y i ) if its output f ( − → x i ) is different from the desired output y i . In other words the weight vector will change only in the case where the model has made an error. The initialization of − → W is usually performed simply by setting − → W (0) = 0.</p><p>The training set T r is said to be linearly separable if there exists a positive constant γ and a weight vector − → W such that</p><formula xml:id="formula_45">y i − → W • − → x i + b &gt; γ, ∀ ( − → x i , y i ) ∈ T r<label>(7)</label></formula><p>Novikoff <ref type="bibr">[4]</ref> proved that the perceptron algorithm converges after a finite number of iterations k if the train data set is linearly separable. The number of mistakes (iterations) is bounded then by</p><formula xml:id="formula_46">k ≤ 2R γ 2<label>(8)</label></formula><p>where R = max{|| − → x i ||} is the maximum norm of an input train vector.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3">Batch Perceptron</head><p>Equation 6 defines a single sample fixed increment perceptron learning rule. It is called fixed increment because parameter α is constant throughout training.</p><p>In the case where this parameter changes at each iteration, we say that it is a variable increment perceptron. It is also called single sample because this rule applies at each instance x i which was misclassified during iteration k. In other</p><formula xml:id="formula_47">words, at iteration k each ( − → x i , y i ) ∈ T r is presented to model − → W (k−1) and if it is misclassified by it (f (k−1) ( − → x i ) = y i ) then this single instance − → x i is used (along with parameter α (k−1) ) to alter − → W (k−1) into − → W (k) .</formula><p>A modification of this perceptron can be made defining a set of instances</p><formula xml:id="formula_48">Err ⊂ T r Err = {( − → x i , y i )}, f (k−1) ( − → x i ) = y i<label>(9)</label></formula><p>that contains all the misclassified examples at iteration k and then modifying weight vector as:</p><formula xml:id="formula_49">− → W (k) = − → W (k−1) + α (k−1) ( − → x i,yi)∈Err y i − → x i<label>(10)</label></formula><p>In the case where bias value is not incorporated into example and weight vectors (via an additional n+1 dimension), then bias value is modified as:</p><formula xml:id="formula_50">b (k) = b (k−1) + α (k−1) ( − → x i,yi)∈Err y i<label>(11)</label></formula><p>Equations 10 and 11 are called a Batch Perceptron learning rule and as the single sample perceptron, parameter α (k−1) can be constant (fixed increment) or varying at each iteration (variable increment).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.4">Centroid Classifier</head><p>A Centroid classifier is a simple linear classifier, that will help us understand the notion behind our modification presented in the next Section. In the simple binary case there are two classes, the positive and the negative one. We define set C + and C − containing instances from the positive and respectively the negative class. We call Centroid of the positive class and respectively the Centroid of the negative class as</p><formula xml:id="formula_51">− → C + = 1 |C + | − → x i ∈C+ − → x i<label>(12)</label></formula><formula xml:id="formula_52">− → C − = 1 |C − | − → x i ∈C− − → x i<label>(13)</label></formula><p>We then define a linear classifier as</p><formula xml:id="formula_53">h : − → W • − → x + b = 0<label>(14)</label></formula><p>where</p><formula xml:id="formula_54">− → W = − → C + − − → C −<label>(15)</label></formula><p>and bias value b is defined by some technique we discuss in the following paragraphs.</p><p>Figure <ref type="figure" target="#fig_0">1</ref> illustrates a simple case of a centroid classifier in a 2-dimensional space. Sets C + of the positive class and C − of the negative class are shown along with their centroid vectors − → C + and − → C − respectively. We note that in this simple example, these two classes are linearly separable and therefore it is possible to find a value for bias b such that h is a perfect separating hyperplane.</p><p>A method for finding such a value is Scut <ref type="bibr">[5]</ref>, where we iteratively choose values for bias b and then keep the one that lead to the best classifier (as measured by some evaluation measurement). Bias takes values as</p><formula xml:id="formula_55">b i = − → W • − → x i , ∀ − → x i ∈ T r<label>(16)</label></formula><p>and then an evaluation measure (for example the F 1 measure) is computed for classifier h :</p><formula xml:id="formula_56">− → W • − → x + b i = 0.</formula><p>Finally as bias value is chosen the one that gave the maximum evaluation measure. It is clear that the instance x i that corresponds to the chosen b i lies on hyperplane h. In the shown 2-dimensional example of Figure <ref type="figure" target="#fig_0">1</ref> this instance is marked by point − → x Scut .</p><p>This simple algorithm has previously investigated and methods have been proposed for altering initial centroids or weights in order to achieve a better classifier <ref type="bibr">[6]</ref><ref type="bibr">[7]</ref><ref type="bibr">[8]</ref>.</p><p>In the next subsection we present how ideas from Centroid Classifier and Perceptron are combined to our modified version of Perceptron. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.5">The proposed modification to Perceptron</head><p>Centroid Classifier of the previous subsection can be seen as a perceptron with initial weight vector</p><formula xml:id="formula_57">− → W (0) = − → C + − − → C − ,</formula><p>bias value b as defined by an Scut method and no other training adjustments at all. The case shown in Figure <ref type="figure" target="#fig_0">1</ref> is an ideal case for a Centroid Classifier, meaning that it is possible to find a value for b resulting to a perfect separating hyperplane h :</p><formula xml:id="formula_58">− → W • − → x + b = 0.</formula><p>This is not however true in all cases. Figure <ref type="figure" target="#fig_4">2</ref> shows such a case where finding a perfect separating hyperplane is not possible for a simple Centroid Classifier. Dark regions contains misclassified instances that cannot correctly classified. A Simple Sample or a Batch Perceptron would use these errors to modify the weight vector − → W . If we define sets F P and F N as:</p><formula xml:id="formula_59">F P = {( − → x i , y i )}∀x i ∈ C − , f ( − → x i ) = y i (<label>17</label></formula><formula xml:id="formula_60">) F N = {( − → x i , y i )}∀x i ∈ C + , f ( − → x i ) = y i (<label>18</label></formula><formula xml:id="formula_61">)</formula><p>in other words set F P contains negative instances that were misclassified as positive (False Positive), whereas set F N contains positive instances that were misclassified as negative (False Negative). A Batch Perceptron then using mis- classified instances modifies weight vector as Equation 10 or equivalently as:</p><formula xml:id="formula_62">− → W (k+1) = − → W (k) + α (k)   − → x i∈F N (k) − → x i − − → x i∈F P (k) − → x i   (<label>19</label></formula><formula xml:id="formula_63">)</formula><p>However there is a parameter α, either constant or variable that needs to be estimated. This learning rate parameter is strongly related to the field on which perceptron learning is applied and train data itself. A way to estimate it is using a validation set of instances and selecting a value for α that leads to maximum performance. But this operation must be repeated whenever field of operation or data is switched and costs very much in terms of time.</p><p>Another approach is to use a fixed value for the learning rate like α = 1 or α = 0.5 for example, without attempting to find a optimal value. However this could result to very unwanted effects because learning rate is too small or too large for the specific field of operation and training instances.</p><p>The key idea of our approach is illustrated in Figure <ref type="figure" target="#fig_10">3</ref> where we concentrate on the misclassified regions. Positive class and a portion of negative class are shown. Initial weight vector − → W (0) and hyperplane h (0) are defined by a simple Centroid Classifier. The idea is, at the next iteration 1, to modify weight vector and bias into − → W (1) and b (1) such that the resulting hyperplane h (1) passes through the points defined by centroid vectors of the misclassified regions F P and F N . We define these misclassified centroids at each iteration as</p><formula xml:id="formula_64">− − → F P (k) = 1 |F P (k) | − → x i ∈F P (k) − → x i<label>(20)</label></formula><formula xml:id="formula_65">− − → F N (k) = 1 |F N (k) | − → x i ∈F N (k) − → x i<label>(21)</label></formula><p>where sets F P and F N are defined in Equations 17 and 18. We then define the error vector at each iteration as</p><formula xml:id="formula_66">− → e (k) = − − → F N (k) − − − → F P (k)<label>(22)</label></formula><p>Batch Perceptron learning rule of Equation 19 is then modified to:</p><formula xml:id="formula_67">− → W (k+1) = − → W (k) + α (k)− → e (k)<label>(23)</label></formula><p>We can easily compute the value of this modified learning rate α (k) if we note that misclassified centroids − − → F N (k) and − − → F P (k) lie by construction on the new hyperplane h (k+1) . As a result error vector − → e (k) is vertical to the new normal</p><formula xml:id="formula_68">vector − → W (k+1) . So − → W (k+1) • − → e (k) = 0 − → W (k) + α (k)− → e (k) • − → e (k) = 0 − → W (k) • − → e (k) + α (k) || − → e (k) || 2 = 0 α (k) = − − → W (k) • − → e (k) || − → e (k) || 2</formula><p>And then the modified learning rule of Equation <ref type="formula" target="#formula_67">23</ref>is</p><formula xml:id="formula_69">− → W (k+1) = − → W (k) − − → W (k) • − → e (k) || − → e (k) || 2 − → e (k)<label>(24)</label></formula><p>This is the normal vector defining the direction of the next hyperplane h (k+1) .</p><p>The actual position of it is determined by the new bias value which is easily computed (bringing in mind that misclassified centroids lie on the new hyperplane):</p><formula xml:id="formula_70">b (k+1) = − − → W (k+1) • − − → F P (k) = − − → W (k+1) • − − → F N (k)<label>(25)</label></formula><p>Equations 24 and 25 define the new hyperplane</p><formula xml:id="formula_71">h (k+1) : − → W (k+1) • − → x + b (k+1) = 0</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Task Description</head><p>As last year's, this year's ECML PKDD Discovery Challenge deals with the well known social bookmarking system called Bibsonomy<ref type="foot" target="#foot_21">1</ref> . In such systems, users can share with everyone links to web pages or scientific publications. The former are called bookmark posts, where the later are called bibtex posts. Apart from posting the link to the page or the publication, users can assign tags (labels) to their posts. Users are free to choose their own tags or the system can assist them by suggesting them the appropriate tags. This year's Discovery Challenge problem is about generating methods that would assist users of social bookmarking systems by recommending them tags for their posts. There are two distinct task for this problem. Task 1 is about recommending tags to posts over an unknown set of tags. That means that the methods developed for Task 1 must be able to suggest tags that are unknown (in other words suggest new tags). Task 2, on the other hand, is about recommending tags that have been already known to be ones. <ref type="foot" target="#foot_22">2</ref></p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Data Description</head><p>Data provided for these tasks was extracted from Bibsonomy databases. Two datasets where provided, one for training participant's methods, and the other for evaluating their performance. Both of them where provided as a set of 3 files (tas, bookmark, bibtex). Files bookmark and bibtex contain textual data of the corresponding posts. File tas contains which user assigned which tags to which bookmark or bibtex resource. Each triplet (user,tags,resource) defines a post. Train and test files where of the same tabular format, except test tas file which of course did not contain tag information, as this was Challenge's goal. <ref type="foot" target="#foot_23">3</ref>More details about preprocessing of the datasets will be given on the following section 4.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Experimental Setup and Results</head><p>Challenge's organizers had suggest that graph method would fit better to task 2, whereas content based method would fit to task 1. In our work we concentrated on task 2, and from this point on whenever we mention a task, we mean task 2. Although organizers suggested graph method for the task, we choose to use our modified perceptron rule for solving this problem. We made this decision because we wanted to test the performance and robustness of the proposed algorithm on a domain with a large category set. As we are going to present in our under review paper, we have evaluated the proposed algorithm on standard text classification datasets as well as on artificially generated (and linearly separable) datasets. Although feature spaces of these datasets are of tens or hundreds of thousands features, their categories sets are of few to at most a thousand categories. We wanted to investigate how this method is going to perform when both feature and category spaces are large.</p><p>So, Task 2 can be seen as a standard text classification problem, and the proposed algorithm as a machine learning, supervised, automatic classification method that applies on it. In this problem, tags (labels) that assigned on posts can be seen as categories. On the other hand, posts can be seen as text documents, where category labels (tags) are assigned on them.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Data Preprocessing</head><p>Viewing task 2 as a supervised text classification problem, implies that datasets must transformed to a vector space, where the proposed linear classifier can be used. For every post (user,tags,resource), we construct a text document and then transform it to the vector space.</p><p>We choose to discard user information from the posts, so the only textual information for each post came from the assigned tags and the resource. Furthermore for each bookmark post we kept url, description and extended description fields. For each bibtex post we kept journal, booktitle, url, description, bibtexAbstract, title and author fields.</p><p>Fore every post, and using those field, we construct a text document. We then transform document dataset to a vector space. First tokenization of the text, then stop word removal, then stemming (using Porter's stemmer <ref type="bibr">[9]</ref>, then term and feature extraction and finally feature weighting using tf*idf statistics.</p><p>The following table <ref type="table" target="#tab_0">1</ref> presents some statistics about categories (tags) and documents (posts) in the train and the test dataset. The following diagram 4 presents the distribution of the sizes of categories in the train dataset. Axis x denotes the number of categories that are of a certain size. Axis y denotes the number of documents that a certain sized category contains. We note that categories sizes are small in general. In fact 10,500 out of 13,276 categories have at most 10 documents. The average size of categories is 1.97 (average posts per tag).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Experimental Setup and Train phase</head><p>After converting documents (posts) into vectors in a high-dimensional space, we can apply the proposed text classification method for solving the multilabel classification problem. Since the method trains a binary linear classifier, the problem must be transformed into binary classification. This is done by cracking the problem into multiple binary classification problems. So, at the end we have to solve 13, 276 binary classification problems.</p><p>The number of problems is quite large and therefore the used method must be as much fast as possible. After the train phase (which finishes after the reasonable time of 2 hours in a mainstream laptop), the final classification system consists of 13, 276 binary classifiers.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">Test phase and Results</head><p>Test phase consists of presenting each document of the test dataset (778 in total) to every binary classifier resulted from training phase (13, 276 in total). Each classifier decides whether the presented document (post) belongs or not to the corresponding category (tag). Time needed fore presenting all document to all classifiers on a mainstream laptop was about 10 minutes (that is about 0.8 seconds for a document to pass through all classifiers).</p><p>We produced 2 types of results. The ones that come from binary classification and the ones that come from ranking. During binary classification a document could be assigned or not into a category. Therefore a document, after been presented to every binary classifier, could be assigned to zero, one, or more categories (max is 13, 276 of course).</p><p>On the ranking mode, a classifier gives a score to each presented document (higher score mean higher confidence of the classifier that this document belongs to the corresponding category). Therefore at this mode, a document can be assigned to any number z of categories we select (simply by selecting the z categories which gave the higher scores).</p><p>We chose our submission to the Challenge, to contain results of the ranking mode (by selecting the 5 higher scored categories for each document).</p><p>After releasing the original tag assignments of the test dataset, our results of the ranking mode achieved a performance of F 1 = 0.1008. The results of the first mode (binary mode), that where never submitted, achieved a performance of F 1 = 0.1622. Of course, those results could not have been known prior releasing original test tas file, but we had a belief that the ranking mode (suggesting 5 tags for every post, instead of less or even zero) would had better results. Unfortunately this belief was false.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Concluding Remarks</head><p>In this paper we described the application of a modified version of the Perceptron learning rule on Task 2 of ECML PKDD Discovery Challenge 2009. This algorithm acts as a supervised machine learning, automatic text classification algorithm on the data of the task. Task 2 is transformed to a supervised text classification problem by treating users' posts ass text documents and assigned tags as thematic categories.</p><p>This algorithm has been prior tested on various text classification datasets and artificially generated linearly separable datasets, and it has shown a robust performance and efficiency. Compared with the original Batch Perceptron learning algorithm, it shows a significant improvement on the convergence rate.</p><p>Its fast training phase made it feasible to be used on Task 2 dataset, which consists of a large categories set (more than 13, 000 categories) and a linear classifier had to be trained for each category.</p><p>Although its results on Task 2 test dataset where not so well, we think that its fast training phase and fast evaluation (since it is just a dot product for each category-document tuple) allow for further investigation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Discriminative Clustering for Content-Based Tag Recommendation in Social Bookmarking Systems</head><p>Malik Tahir Hassan 1 , Asim Karim 1 , Suresh Manandhar 2 , and James Cussens  The University of York York, UK {mhassan, akarim}@lums.edu.pk; {suresh, jc}@cs.york.ac.uk Abstract. We describe and evaluate a discriminative clustering approach for content-based tag recommendation in social bookmarking systems. Our approach uses a novel and efficient discriminative clustering method that groups posts based on the textual contents of the posts. The method also generates a ranked list of discriminating terms for each cluster. We apply the clustering method to build two clustering models -one based on the tags assigned to posts and the other based on the content terms of posts. Given a new posting, a ranked list of tags and content terms is determined from the clustering models. The final tag recommendation is based on these ranked lists. If the poster's tagging history is available then this is also utilized in the final tag recommendation. The approach is evaluated on data from BibSonomy, a social bookmarking system. Prediction results show that the tag-based clustering model is more accurate than the termbased clustering model. Combining the predictions from both models is better than either model's predictions. Significant improvement in recommendation is obtained over the baseline method of recommending the most frequent tags for all posts.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Social bookmarking systems have become popular in recent years for organizing and sharing resources on the Web. Such systems allow users to build a database of resources, typically Web pages and publications, by adding basic information (such as URLs and titles) about them and by assigning one or more keywords or tags describing them. The tags serve to organize the resources and help improve recall in searches. Individual users' databases are shared among all users of the system enabling the development of an information repository which is commonly referred to as a folksonomy <ref type="bibr">[1]</ref>. A folksonomy is a collection of users, resources, and tags assigned by a user to a resource posted by him or her. Tag recommendation for new posts by users is desirable for two reasons. First, it ensures uniformity of tagging enabling better searches, and second, it eases the task of users in selecting the most descriptive keywords for tagging the resource.</p><p>Tag recommendation can have one of two goals: (1) to suggest tags tailored to individual users' preferences (the 'local' goal). and (2) to suggest tags that promote uniformity in tagging of resources (the 'global' goal). Tag recommendation can benefit from the tagging history of users and resources. However, when a user posts for the first time and/or the posted resource is new this historical information is less useful. In such cases, content-based tag recommendation is necessary, in which the contents of the resource are relied upon for tag recommendation.</p><p>This paper addresses task 1 of the ECML PKDD Discovery Challenge 2009 <ref type="bibr">[2]</ref>. This task deals with content-based tag recommendation in BibSonomy, a social bookmarking system. The goal of tag recommendation is 'local', that is, to suggest tags tailored to individual users' preferences. Historical data of users, resources, and tags is available; however, the tag recommendation system must be able to provide good recommendations for unseen users and/or resources. Thus, the contents of resources must be utilized for tag recommendation.</p><p>Our solution to task 1 of the ECML PKDD Discovery Challenge 2009 relies on a novel discriminative clustering and term ranking method for textual data. We cluster the historical data of posted resources and develop a ranked list of discriminating tags and content terms for each cluster. Given a new posting, based on its contents, we find the best 3 clusters and develop a weighted list of tags and terms appropriate for tagging the post. If the poster's tagging history is available, then this provides a third ranked list of tags appropriate for the post. The final tag recommendation for the post is done by rules that select terms from the weighted lists. These rules also decide on the number of tags to recommend for each known poster. Extensive performance results are presented for the post-core training data provided by the challenge organizers.</p><p>The rest of the paper is organized as follows. We present the related work and motivation in Section 2. Section 3 presents details of our content-based tag recommendation approach, including description of the discriminative clustering and method. Data preprocessing and analysis is discussed in Section 4. The results of our approach are presented and discussed in Section 5. We conclude in Section 6.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Related Work and Motivation</head><p>Tagging resources with one or more words or terms is a common way of organizing, sharing, and indexing information. Tagging has been popularized by Web applications like image (e.g flickr), video (e.g. YouTube), bookmark (e.g. dec.icio.us), and publication (e.g. BibSonomy) sharing/organizing systems. Automatic tag recommendation for these applications can improve the organization of the information through 'purposeful' tag recommendations. Moreover, automatic tag recommendations ease the task of users while posting new resources.</p><p>The majority of the approaches proposed for tag recommendation assume that either the user posting the resource and/or the resource itself has been seen in the historical data available to the system <ref type="bibr">[3]</ref><ref type="bibr">[4]</ref><ref type="bibr">[5]</ref><ref type="bibr">[6]</ref>. If this is not the case, then only the contents of the posted resource can be relied upon. For social bookmarking systems, contents of resources are textual in nature requiring appropriate text and natural language processing techniques.</p><p>Content-based tag recommenders for social bookmarking systems have been proposed by <ref type="bibr">[7,</ref><ref type="bibr">8]</ref>. Lipczak's method extracts the terms in the title of a post, expands this set by using a tag co-occurrence database, and then filters the result by the poster's tagging history <ref type="bibr">[7]</ref>. He reports significant improvements in performance after each step of this three step process. Tatu et al.'s method utilizes terms from several fields including URL and title to build post and user based models <ref type="bibr">[8]</ref> . It relies on natural language processing to normalize terms from various sets before recommending them. We use terms from several fields of the posts including URL and title. We also study the impact of filling in missing and augmenting fields from information crawled from the Web.</p><p>A key challenge in tag recommendation is dealing with sparsity of information. In a typical collaborative tagging system, the vast majority of tags are used very infrequently making learning tagging behavior very difficult. This issue is often sidestepped in evaluation of tag recommenders when they are evaluated on post-core data with a high level of duplication (e.g. in <ref type="bibr">[4,</ref><ref type="bibr">6]</ref> post-core at level 5 is used). Our evaluation is done on post-core at level 2 data provided by the ECML PKDD Discovery Challenge 2009 <ref type="bibr">[2]</ref>.</p><p>Document clustering has been used extensively for organizing and summarizing large document collections <ref type="bibr">[9,</ref><ref type="bibr">10]</ref>. A useful characteristic of clustering is that it can handle sparse document spaces by identifying cohesive groups. However, clustering is generally computationally expensive. In the domain of collaborative tagging systems, clustering has been explored for information retrieval and post recommendation <ref type="bibr">[11,</ref><ref type="bibr">12]</ref>. In this paper, we explore the use of clustering for content-based tag recommendation. We use an efficient method that is practical for large data sets.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Discriminative Clustering for Content Based Tag Recommendation</head><p>Our approach for content-based tag recommendation in social bookmarking systems is based on discriminative clustering, content terms and tags rankings, and rules for final recommendations. We use a novel and efficient discriminative clustering method to group posts based on the tags assigned to them and based on their contents' terms. This method maximizes the sum of the discrimination information provided by posts and outputs a weighted list of discriminating tags and terms for each cluster. We also maintain a ranked list of tags for seen users. Tags are suggested from these three rankings by intuitive rules that fuse the information from the lists. The rest of this section presents our approach in detail.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Problem Definition and Notation</head><p>A social bookmarking system, such as BibSonomy <ref type="bibr">[13]</ref>, allows users to post and tag two kind of resources: Web bookmarks and publications. Each resource type is described by a fixed set of textual fields. A bookmark is described by fields like URL, title, and description, while a publication is described by fields in the standard bibtex record. Some of these fields (like title for bookmarks) are mandatory while others are optional. This textual information forms the content of the resource. Each user who posts a resource must also assign one or more tags for describing the resource. Let p i = {u i , x i , t i } denotes the ith post, where u i is the unique user/poster ID, and x i and t i are the vector space representations of the post's contents and tags, respectively. If T is the size of the vocabulary then the ith post's contents and tags can be written as x i = {x i1 , x i2 , . . . , x iT } and t i = {t i1 , t i2 , . . . , t iT }, respectively, where x ij (t ij ) denotes the frequency of term j (tag j) in post i. Note that an identical vector space model is used to represent both content terms and tags, t ij ∈ {0, 1}, ∀i, j, and x ij ≥ 0, ∀i, j. The historical data contain N posts. The tag recommender suggests tags for a new post i described by u i and x i . The user u i and resource described by content x i may or may not appear in the historical data.</p><p>Let T G(i), T M (i), and T U (i) be the ranked list of tags from clustering, terms from clustering, and user tags, respectively, corresponding to the ith post. The actual tags recommended for post i, denoted by T R(i), are determined from these ranked lists by intuitive rules.</p><p>Given a test data containing M posts, the performance of the tag recommender is evaluated by averaging F1-score of each prediction over the entire test data.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Discriminative Clustering for Tag and Term Ranking</head><p>The historical data of N posts is clustered into K N groups using a novel discriminative clustering method. This method is motivated from the recently proposed DTWC algorithm for text classification <ref type="bibr">[14]</ref>. It is an iterative partitioning method that maximizes the sum of discrimination information provided by each textual content (a post, in our setting) between its assigned cluster and the remaining clusters. The key ideas include discriminative term weighting, discrimination information pooling, and discriminative assignment. Unlike other partitioning clustering methods, this method does not require the explicit definition of a similarity measure and a cluster representative. Furthermore, it builds a ranked list of discriminating terms for each cluster implicitly. The method is computationally more efficient than popular methods like the k-means clustering algorithm. We perform two clusterings of the historical data -one based on the content terms x and the other based on the tags t of the posts in the data. In the following description, we develop the method for content terms only; the method as applied to tags will be similar.</p><p>First, an initial clustering of the data is done. This can be done randomly or, less efficiently especially for large collections, by a single iteration of the k-means algorithm with the cosine similarity measure. Given this clustering, a discriminative term weight w k j is computed for each term j in the vocabulary and for each cluster k as [14]</p><formula xml:id="formula_72">w k j = p(x j |k)/p(x j |¬k) when p(x j |k) &gt; p(x j |¬k) p(x j |¬k)/p(x j |k) otherwise</formula><p>where p(x j |k) and p(x j |¬k) are the probabilities that term j belongs to cluster k and the remaining clusters (¬k), respectively. The discriminative term weight quantifies the discrimination information that term j provides for cluster k over the remaining clusters. Note that this weight is expressed as a probability ratio and is always greater than or equal to 1. The probabilities are computed by maximum likelihood estimation from the historical data.</p><p>Having computed the discriminative term weights for the current clustering, two discrimination scores can be computed for each post i. One score, denoted as Score k (x i ), expresses the discrimination information provided by post i for cluster k, whereas the other score, denoted as Score ¬k (x i ), expresses the discrimination information provided by post i for clusters ¬k. These scores are computed by linearly pooling the discrimination information provided by each term x j in post i as <ref type="bibr">[14]</ref> Score k (x i ) = j∈Z k x j w k j j x j and Score ¬k (x i ) = j∈Z ¬k x j w k j j x j In these equations, Z k = {j|p(x j |k) &gt; p(x j |¬k)} and Z ¬k = {j|p(x j |¬k) &gt; p(x j |k)} are sets of term indices that vouch for clusters k and ¬k, respectively. Each post, described by its contents x, is then reassigned to the cluster k for which the cluster score f k = Score k (x) − Score ¬k (x) is maximum. This is the cluster that makes each post most discriminating among all the clusters.</p><p>The overall clustering objective is to maximize the sum of discrimination information, or cluster scores, of all posts. Mathematically, this is written as</p><formula xml:id="formula_73">Maximize J = N i=1 K k=1 I k (x i ) • f k where I k (x i ) = 1 if post i</formula><p>is assigned to cluster k and zero otherwise. Iterative reassignment is continued until the change in the clustering objective becomes less than a specified small value. Typically, the method converges satisfactorily in fewer than 15 iterations.</p><p>The discriminative term weights for the terms in the index set Z k are ranked to obtain the weighted and ranked list of terms for cluster k. As mentioned earlier, clustering is also performed based on the tags assigned to posts. This clustering yields another weighted and ranked list of tags for each cluster.</p><p>It is worthwhile to point out that the term-based clustering is done on both the training and testing data sets. This approach allows the terms that exist only in the test data to be included in the vocabulary space, and for such terms to be available for recommendation as tags.</p><p>Given a new post i described by x i , the best cluster for it is the cluster k for which the cluster score f k is a maximum. The corresponding ranked list of terms and tags for post i are denoted by T M (i) and T G(i), respectively. These ranked lists contain the most discriminating tags for post i based on its contents.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Final Tag Recommendation</head><p>Given a new post, and based on the contents x of the post, two ranked lists of terms appropriate for tagging are generated by the procedures described in the previous section.</p><p>If the user of the post appears in the historical data, then an additional list of potential tags can be generated. This is the ranked list of tags T U (i) used by the user of post i The ranking is done based on frequency. Moreover, the average number of tags per user is computed and used while recommending tags for seen users.</p><p>The final list of tags for post i is made by simple and intuitive rules that combine information from all the lists. Let S be the number of tags to recommend for post i. Then, the final list of tags for the post is given by the following algorithm:</p><formula xml:id="formula_74">T R(i) = T G(i)[1 : P ] ∩ T M (i)[1 : Q] IF T U (i)| = THEN T R(i) = T R(i) ∩ T U (i)[1 : R] IF |T R(i)| &lt; S THEN add top terms from T G(i), T M (i)inT R(i)</formula><p>In the above algorithm, P , Q, and R are integer parameters that define how many top terms to include from each list. If after taking the set intersections |T R(i)| &lt; S then the remaining tags are obtained from the top tags and terms in T G(i) and T M (i), respectively. In general, as seen from our evaluations, R ≤ Q ≤ P , indicating that T G(i) is the least noisy source and T U (i) the most noisy source for tags.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Evaluation Setup</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Data and their Characteristics</head><p>We evaluate our approach on data sets made available by the ECML PKDD Discovery Challenge 2009 <ref type="bibr">[2]</ref>. These data sets are obtained from dumps of public bookmark and publication posts on BibSonomy <ref type="bibr">[13]</ref>. The dumps are cleaned by removing spammers' posts and posts from the user dblp (a mirror of the DBLP Computer Science Bibliography). Furthermore, all characters from tags that are neither numbers nor letters are removed. UTF-8 encoding and unicode normalization to normal form KC are also performed.</p><p>The post-core at level 2 data is obtained from the cleaned dump (until 31 December 2008) and contain all posts whose user, resource, and tags appear in at least one more post in the post-core data. The post-core at level 2 contain 64,120 posts (41,268 bookmarks and 22,852 publications), 1,185 distinct users, and 13,276 distinct tags. We use the first 57,000 posts (in content ID order) for training and the remaining 7,120 posts for testing.</p><p>We also present results on the test data released as part of task 1 of the ECML PKDD Discovery Challenge 2009. This data is cleaned and processed as described above, but it contain only those posts whose user, resource, or tags do not appear in the post-core at level 2 data. This data contain 43,002 posts (16,898 bookmarks and 26,104 publications) and 1,591 distinct users. For this evaluation, we use the entire 64,120 posts in the post-core at level 2 for training and test on the 43,002 posts in the test data.</p><p>These data sets are available in the form of 3 tables -tas, bookmark, and bibtexas described below. The content of a post is defined by the fields in the bookmark and bibtex tables, while the tags appear in the tas table.</p><p>tas fact table; who attached which tag to which post/content. Fields include: user (number; user names are anonymized), tag, content id (matches bookmark.content id or bibtex.content id), content type (1 = bookmark, 2 = bibtex), date bookmark dimension table for bookmark data. Fields include: content id (matches tas.content id), url hash (the URL as md5 hash), url, description, extended description, date bibtex dimension table for BibTeX data. Fields include: content id (matches tas.content id), journal, volume, chapter, edition, month, day, booktitle, howPublished, institution, organization, publisher, address, school, series, bibtexKey (the bibtex key (in the @... line)), url, type, description, annote, note, pages, bKey (the "key" field), number, crossref, misc, bibtexAbstract, simhash0 (hash for duplicate detection within a user -strict -(obsolete)), simhash1 (hash for duplicate detection among users -sloppy -), simhash2 (hash for duplicate detection within a userstrict -), entrytype, title, author, editor, year A few tagging statistics from the post-core data are given in Table <ref type="table" target="#tab_0">1</ref> and Figure <ref type="figure" target="#fig_0">1</ref>. These statistics are used to fix the parameter S (number of recommended tags) for known users. For unseen users, S is set at 5. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Data Preparation</head><p>We explore tag recommendation performance on original contents, contents that have been augmented by crawled information, and contents that have been augmented and lemmatized. The vocabulary for the vector space representation is formed from the tags and content terms in the training and testing sets. Selected content fields are used for gathering the content terms. For bookmark posts, the selected fields are url, description, and extended. For publication posts, the selected bibtex fields are booktitle, journal, howpublished, publisher, series, bibtexkey, url, description, annote, note, bkey, crossref, misc, bibtexAbstract, entrytype, title, and author. As mentioned earlier, the tags, which appear in the tas table, are also included in the vocabulary.</p><p>We remove all the non-letter and non-digit characters, but retain umlauts and other non-Latin characters due to UTF-8 encoding. All processed terms of length greater than or equal to three are retained. The tags are processed similarly, but without considering the token length constraint. Crawling Crawling is done to fill in and augment important fields. For bookmark posts, the extended description field is appended with textual information from &lt;TITLE&gt;, &lt;H1&gt; and &lt;H2&gt; HTML fields of the URL provided in the posts.</p><p>For publication posts, missing abstract field are filled using online search. We use the publication title to search for its abstract on CiteULike <ref type="bibr">[15]</ref>. If the article is found, and its abstract is available on CiteULike, the bibtexAbstract field of the post is updated. CiteULike is selected because its structure is simpler and it does not have any restrictions on the number of queries (in a day for example).</p><p>Lemmatization We also explore lemmatization of the vocabulary while developing the vector space representation. Lemmatization is different from stemming as lemmatization returns the base form of a word rather than truncating it. We do lemmatization using TreeTagger <ref type="bibr">[16]</ref>. TreeTagger is capable of handling multiple languages besides English. We lemmatize the vocabulary using English, French, German, Italian, Spanish and Dutch languages. The procedure, in brief, is as below: 4. If a word is lemmatized by more than one language, then lemmas are prioritized in the sequence: English, French, German, Italian, Spanish, Dutch. The first lemma for the word is selected.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">Evaluation Criteria</head><p>The performance of tag recommendation systems is typically evaluated using precision, recall, and F1 score, where the F1 score is a single value obtained by combining both precision and recall. We report the precision, recall, and F1 score averaged over all the posts in the testing set.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Results</head><p>In this section, we present and discuss the results of our discriminative clustering approach for content based tag recommendation. We start off by evaluating the performance of the clustering method.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1">Clustering Performance</head><p>The performance of the discriminative clustering method is evaluated on the entire 64,120 posts of the post-core at level 2 data. We cluster these posts based on the tags assigned to them. After clustering and ranking of tags for each cluster, we recommend the top 5 tags from the ranked list for all posts in each cluster. The average precision, recall, and F1 score percentages obtained for different values of K (number of desired clusters) is shown in Table <ref type="table" target="#tab_1">2</ref>.</p><p>The top 5 tags become increasingly accurate recommendations as the number of clusters is increased, with the maximum recall of 48.7% and F1 score of 30.6% obtained when K = 300. These results simulate the scenario when the entire tag space (containing 13,276 tags) is known. Furthermore, there is no separation between training and testing data. Nonetheless, the results do highlight the worth of clustering in grouping related posts that can be tagged similarly. Table <ref type="table" target="#tab_3">3</ref> shows the top ranked tags for selected clusters. It is seen that the discriminative clustering method is capable of grouping posts and identifying descriptive tags for  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2">Tag Recommendation Using TG and TM Only</head><p>In this section, we discuss the performance of recommending the top 5 tags from the T G(i) or T M (i) list of each post i. This evaluation is done on the testing data of 7,120 posts held out from the post-core at level 2 data. The clustering model is based on the first 57,000 posts (in content ID order) from the data. In this evaluation, the original data, without augmentation with crawled information, is used for creating the vector space representation.</p><p>The recommendation results for different K values are given in Table <ref type="table" target="#tab_5">4</ref>. Results are shown for the case when only the top cluster for each post is considered, and for the case when the top three clusters of each post are merged in a weighted manner (using cluster score and discriminative term weights). It is observed that merging the lists of the top three clusters always gives better performance. Moreover, recommendations based on T G(i) are always better than those based on T M (i) indicating that the term-based clustering is more noisy than that based on tags. We also find out that K = 200 yields the highest recommendation performances. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3">Tag Recommendation Using All Lists</head><p>In this section, we evaluate the performance of our approach when utilizing information from all lists. We also evaluate performance on original, crawled, and crawled plus lemmatized data. These results are shown in Table <ref type="table" target="#tab_7">5</ref>. For this evaluation, we fix K = 200 and use the top three clusters for building T G(i) and T M (i).</p><p>The first column (identified by the heading TF) shows the baseline result of recommending the top 5 most frequent tags in the training data (57,000 posts from post-core data). It is seen that our clustering based recommendation improves performance beyond the baseline performance. The second and third columns show the performance of recommending the top 5 terms from T G(i) and T M (i), respectively. The predictions of the tag-based clustering always outperform the predictions of the term-based clustering. In the fourth column, we report results for the case when the top 5 recommended tags are obtained by combining T G(i) and T M (i), as described in Section 3.3. These results are significantly better than those produced by each list independently.</p><p>The fifth column shows the results of combining all lists, including the user list T U (i) when known. This strategy produces the best F1 score of 15.5% for the crawled data. This is a significant improvement over the baseline F1 score of 7.0%.</p><p>Table <ref type="table" target="#tab_7">5</ref> also shows that filling in missing fields and augmenting the fields with crawled information improves performance. Lemmatization does not help, probably because users do not necessarily assign base forms of words as tags. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.4">Tag Recommendation for Task 1 Test Data</head><p>We report the performance of our approach on task 1 test data released by the challenge organizers on the bottom line of Table <ref type="table" target="#tab_7">5</ref>. We filled in missing and augmented other fields by crawled information. No lemmatization is done. The final vocabulary size is equal to 317,283 terms making the tag recommendation problem very sparse.</p><p>The baseline performance of using the 5 most frequent tags from the post-core at level 2 (the training data for this evaluation) is the F1 score of 1.1% only. By using our discriminative clustering approach, the average F1 score reaches up to 5.4%. This low value is attributable to the sparseness of the data, and it is unlikely that other methods can cope better without extensive semantic normalization and micro modeling of the tagging process.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Conclusion</head><p>In this paper, we explore a discriminative clustering approach for content-based tag recommendation in social bookmarking systems. We perform two clusterings of the posts: one based on the tags assigned to the posts and the second based on the content terms of the posts. The clustering method produces ranked lists of tags and terms for each cluster.</p><p>The final recommendation is done by using both lists, together with the user's tagging history if available. Our approach produces significantly better recommendations than the baseline recommendation of most frequent tags.</p><p>In the future, we would like to explore language specific models, incorporation of a tag extractor method, and semantic relatedness and normalization.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Time based Tag Recommendation using Direct and Extended Users Sets</head><p>Tereza Iofciu and Gianluca Demartini</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>L3S Research Center Leibniz Universität Hannover</head><p>Appelstrasse 9a D-30167 Hannover, Germany {iofciu,demartini}@L3S.de</p><p>Abstract. Tagging resources on the Web is a popular activity of standard users. Tag recommendations can help such users assign proper tags and automatically extend the number of annotations available in order to improve, for example, retrieval effectiveness for annotated resources.</p><p>In this paper we focus on the application of an algorithm designed for Entity Retrieval in the Wikipedia setting. We show how it is possible to map the hyperlink and category structure of Wikipedia to the social tagging setting. The main contribution is a time-based methodology for recommending tags exploiting the structure in the dataset without knowledge about the content of the resources.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Tagging Web resources has become a popular activity mainly due to the availability of tools and systems making it easy to tag and also due to the advantage users see in tagging their resources. People can for example get better search results, or they can get new resources recommended based on tags other people assigned. One particular problem is the one of recommending relevant tags to users for resources they have introduced in the system. Being able to effectively recommend tags would, firstly, simplify the tasks of the users on the web who want to tag resources (e.g., bookmarks, pictures, . . . ), and, secondly, would allow an automatic annotation of resources that enables, for example, a better search for resources or an improved resource recommendation.</p><p>When we want to assign a tag to a resource (or, to predict which tag a user would assign to a resource) a possible approach is to use the most popular tags for the given resource of the given user. Of course, this is not working well because users can tag resources which are different and people tag the same resource in different ways. For this reason most effective approaches look at the content of the resources and perform more complex analysis of the structure connecting users, resources, and tags.</p><p>Previous approaches focus on the content for resources (e.g., textual content of a web page) or on the structure of the tripartite graph composed of users, resources, and tags. The approaches we propose in this paper do not take into account the content of the resources but only the connection structure in the graph. Additionally, we put more importance on more recent tags with the assumption that users' interests might change over time.</p><p>We adapt an algorithm proposed for ranking entities in Wikipedia <ref type="bibr">[1]</ref> based on a set of initial relevant examples (e.g., already tagged resources) and on the structure of hyperlinks connecting pages and categories containing them. As we defined hard links between documents and categories they belong to and soft links between documents and categories containing linked documents, so we define these types of links between resources/users and tags in the tag recommendation setting.</p><p>The rest of the paper is structured as follows. In Section 2 we describe the proposed algorithms also showing the correspondence to the Wikipedia setting. In Section 3 we describe the experimental setting and results. In Section 4 we compare our work with previously proposed approaches and, finally, in Section 5 we conclude the paper.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Graph Based Algorithms</head><p>In this section we describe the algorithms we designed and used for the graph based task that have been run at Discovery Challenge (DC) 2009.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Using the Resource-User Graph</head><p>In both submitted approaches, starting from the input query post (i.e., the posts from the test file) we retrieve the resource it refers to. We call this resource the query resource. For the query resource we retrieve, using the train data, all the users that have annotated it in different posts. We call this set of users the direct user set. We then use this set of user as an input for the algorithm and retrieve all tags the users have assigned. In the second algorithm, in addition to the set of direct users, we also retrieve the user neighborhood (i.e., users that used at least once a tag in common with the given user). We then use the reunion of the two user sets as input for recommending tags. We call the reunion of the two user sets, the extended user set. As a third approach we have also retrieved just the tags that have previously been assigned to the resource as baseline for comparison.</p><p>As seen in Figure <ref type="figure" target="#fig_0">1</ref>, by traversing the post -resource -users graph, we obtain the set of direct users that have annotated the resource given in the query post. The extended user set is obtained by adding also the neighborhood users to the direct user set, see Figure <ref type="figure" target="#fig_4">2</ref>. We considered two users as being neighbors if they had common tags.</p><p>As a baseline approach we considered the recommendation of the most popular tags for a resource, where we only kept the tags assigned by the direct users to the resource of the query post.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Comparison to the Wikipedia scenario</head><p>The algorithms described in this paper are adapted from those developed for finding relevant results for Entity Retrieval queries in the Wikipedia Setting <ref type="bibr">[1]</ref>. This work was performed in the context of the Entity Ranking track at the evaluation initiative INEX 2008 <ref type="bibr">[2]</ref>. In the following we describe how we can map the Entity Ranking setting with the tag recommendation one.</p><p>In the Wikipedia setting we have as input a set of example entities. The goal is to extend such set with other relevant entities. If, for example, the initial set for the query "European Countries" contains Italy, Germany, and France, then the goal is to extend this list with entities such as Spain, Slovenia, Portugal, . . . Our approach is to retrieve other entities based on common assigned Wikipedia categories. We extract two sets of categories, hard categories as direct categories (similarly to the direct user set) and soft categories from the neighboring entities (i.e., following hyperlinks between Wikipedia articles). As neighboring entities we considered the most frequent entities the example entities linked to (similarly to the extended user set). In the Wikipedia setting entities link to entities via hyperlinks, and each entity has several categories assigned to it.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3">Time dependent tag ranking</head><p>Following the intuition that tags can get outdated over the years, and, thus, older assigned tags should be weighted less for recommendation, we introduced a time decaying function of posts. Scores are assigned to posts based on the time when they have been issued compared to the time the latest test post has been issued. The time decaying function is defined by the following formula:</p><formula xml:id="formula_75">postScore i = λ ∆T imei (1)</formula><p>with the decaying factor lambda being smaller than 1 and the time difference being calculated in years. The tag scores are computed based on the tag specificity (i.e., how often they have been assigned) defined as:</p><formula xml:id="formula_76">tagSpecificity i = log(50 + tagCount i )<label>(2)</label></formula><p>Given the different user sets for a query post, we extract from the training data the most frequent common tags the users have assigned. The tag score is computed based on the formula:</p><formula xml:id="formula_77">tagScore i = j (postScore j ) tagSpecif icity i<label>(3)</label></formula><p>where a post j was considered only if it was posted by one of the users from the direct user set for the first approach and from the extended user set for the second approach. The tags are sorted based on this score and the top five tags are kept and recommended.</p><p>As a baseline, we ranked the tags based on popularity within the resource (i.e., how often a tag has been assigned to a resource) also keeping into account when they had been assigned to the resource, based on the formula:</p><formula xml:id="formula_78">tagScore i = j (postScore j ) (4)</formula><p>3 Experiments</p><p>Experiments were performed on the DC 2009 benchmark<ref type="foot" target="#foot_24">1</ref> in order to evaluate the proposed algorithms.</p><p>Starting from the query posts in the test file we recommended for each post the top five tags using the two described approaches and the baseline. In Figure <ref type="figure" target="#fig_10">3</ref> it is possible to see effectiveness values for the two approaches when a different number of retrieved tags is considered. We can see that the direct user approach performs better. Figure <ref type="figure" target="#fig_15">4</ref> shows the same result with Precision/Recall curves of the two proposed approaches. In Figures <ref type="figure" target="#fig_24">5 and 6</ref> we measure the impact of using the time information when recommending the most popular tags for a resource. With a value of 0.9 for λ, in the time decaying function, the scores were slightly lower than when using just the popularity information (Figure <ref type="figure" target="#fig_23">5</ref>). When using a value of 0.95 for λ, there is a small improvement over the baseline when considering 4 and 5 tags (see Figure <ref type="figure" target="#fig_24">6</ref>). We ran experiments also with values smaller than 0.9 for λ which have shown that Precision and F-measure decrease quite a lot (3% for F-measure with λ = 0.1).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Related Work</head><p>Previous work on tag recommendation mainly distinguish between those looking at the content of the resources and those looking at the structure connecting users, resources, and tags.</p><p>Approaches looking at content of resources for tag recommendations are, for example, <ref type="bibr">[5]</ref> which looks at content-based filtering techniques. In <ref type="bibr">[6]</ref> the authors also look at collaborative tag suggestion in order to identify most appropriate tags.</p><p>A specific area of this field looks at recommending tags focusing on an individual user rather than providing general recommendation for a resource. In <ref type="bibr">[4]</ref> they first create a set of candidate tags to be recommended and then they filer it based on the previous tag a particular user has assigned in the past. In <ref type="bibr">[3]</ref> the FolkRank algorithm is evaluated and compared with simpler approaches. This is a graph based approach that computes popularity scores for resources, users, and tags based on the well-known PageRank algorithm exploiting the link structure. The assumption is that resources which are tagged with important tags by important users becomes important themself. Similarly to FolkRank, our approach exploits the link structure between users, resources, and tags, but rather looks at the vicinity of a post (i.e., a [resources,user] pair) in order to compute a weight for the most appropriate tags.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Conclusions and Further Work</head><p>In this paper we presented our first approaches for tag recommendation using graph information. We proposed two approaches, where, given a query post, we retrieve two sets of users. Based on the tags assigned by users in these sets we recommend new tags. The first set of users, the direct user set, consists of the users that have tagged the resource referred to by the query post. The second set of user, the extended user set, consists of the direct user set as well as the users who are neighbours based on commonly assigned tags to the users in the direct set. The tag scores have been computed keeping into account also the time when they have been assigned. With the proposed approaches, we evaluated the effect of the tag posting time. We compared a time dependent ranking to a tag popularity. In the future, we aim at giving a higher importance to the user given in the query post than to the rest of the direct users. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Social bookmarking systems such as BibSonomy<ref type="foot" target="#foot_27">1</ref> and Delicious<ref type="foot" target="#foot_28">2</ref> have increasingly been used for sharing bookmarking information on the Web resource. Such systems are generally built on a set of collectively-annotated informal tags, comprising a folksonomy. A tag recommendation system could guide users during the bookmarking procedure by providing a suitable set of tags for a given resource. In this paper, we propose a simple but effective approach for tackling the tag recommendation problem. The gist of our method is to appropriately combine different information sources with pre-elimination of barely-used tags.</p><p>The candidate tags for recommendation can be extracted from the following information sources. First, resources themselves may have the annotated tags. For example, the title of a journal article is likely to include some of the annotated keywords. Second, the tags previously annotated by other users for the same resource could be a good candidate set. Third, previously annotated tags for other resources by the same user could also provide some information. <ref type="foot" target="#foot_29">3</ref>The paper is organized as follows. In Section 2, the proposed tag recommendation method is detailed. Then, Section 3 shows the results of experimental evaluation on the training dataset, confirming the effectiveness of the proposed method. Performance of our method on the test dataset is briefly described in Section 4. Finally, concluding remarks are drawn in Section 5.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">The Method</head><p>In this section, we detail the proposed tag recommendation method. First, the procedure for keyword extraction from resource descriptions with importance estimation and filtering is explained. Then, the keyword extraction and importance estimation method from previously annotated information is described. Finally, tag recommendation by combining multiple information sources is explained.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Keyword Extraction from Documents (Resource Descriptions)</head><p>In our approach, candidate keywords are extracted from the columns url, description, and extended description of the table bookmark as well as the columns journal, booktitle, description, and title of the table bibtex. It should be noted here that the candidates extracted from different fields are processed separately. This means that even the same keywords could have multiple importance values according to the columns from which they are extracted. <ref type="foot" target="#foot_30">4</ref>In order to estimate the importance of each keyword, its accuracy and frequency ratios are calculated as follows. </p><formula xml:id="formula_79">Accuracy Ratio, AR(k) = ∑ MC(, ) ∈ ∑ EC(, ) ∈ ⁄ . (1) Frequency Ratio, FR(k) = ∑ EC(, ) ∈ ∑ TEC() ∈ ⁄ .<label>(2)</label></formula><p>The accuracy and frequency ratios of each keyword are calculated across all the documents.</p><p>The keywords whose accuracy is lower than average are not considered for recommendation. This elimination procedure is implemented by the following criterion, which also penalizes frequent words. The keywords in Table <ref type="table" target="#tab_0">1</ref> have accuracy ratios much higher than the average, satisfying Equation (3). In Table <ref type="table" target="#tab_1">2</ref>, we present some keywords on the border with respect to Limit Condition. </p><formula xml:id="formula_80">(k) = EC(k, d)ⅹAR(k). (<label>4</label></formula><formula xml:id="formula_81">)</formula><p>The accuracy weight, AW DS (k), is calculated when recommending tags for a given document (resource) d.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Keyword Extraction from Previously-Annotated Information</head><p>Candidate keywords could be extracted from the previously annotated tags for the same resource. For the BibTex references, the field simhash1 of the table bibtex is adopted for the semantically-same resource detection. For the bookmarks, a pruning function, which has similar effect of the approach used in <ref type="bibr">[2]</ref>, was implemented and deployed in our experiments. These candidate keywords are stored in r-keyword set (RS). Their accuracy weight is calculated as follows. </p><p>Candidate keywords are also extracted from the previously annotated tags by the same person. These candidate keywords are stored in u-keyword set (US). Their accuracy weight is obtained as follows.</p><p>D: set of all documents (resources) which are previously tagged by user u. UC(k, d):</p><formula xml:id="formula_83">1 if document d has keyword k; 0 otherwise. Accuracy Weight from User Set, AW US (k) = ∑ UC(, ). ∈<label>(6)</label></formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3">Tag Recommendation by Combining Multiple Information Sources</head><p>The last step is to recommend appropriate tags from the three candidate keyword sets, i.e., d-keyword set (DS), r-keyword set (RS), and u-keyword set (US). Given a specific user and a document (resource) for tagging, these three candidate keyword sets are specified with accuracy weight for each candidate. Before unifying these candidates, the accuracy weights are normalized into [0, 1] as follows. The above four factors are linearly combined with appropriate coefficients. We have experimented with different coefficient values results. First, we focused on the fact that the d-keyword set (DS) is higher than that from Figure <ref type="figure" target="#fig_0">1</ref> compares the performance when the number of recommended tags We also added tag frequency information, denoting how many times a tag was annotated during the training period. This tag frequency rate is calculated as follows.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>EK</head><formula xml:id="formula_84">TagCount( , ) ∑ ∑ TagCount ∈ ∈ ⁄ (, ); 0 ≤ TFR(</formula><p>) denotes the number of occurrences of a tag t annotated for D denote the set of all tags and the set of all documents (resources), respectively.</p><p>The above four factors are linearly combined with appropriate coefficients. We have experimented with different coefficient values, trying to obtain nearly First, we focused on the fact that the performance of extracted keywords from ) is higher than that from r-keyword or u-keyword sets (RS Figure <ref type="figure" target="#fig_0">1</ref> compares the performance using each keyword set on the training dataset when the number of recommended tags is five.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>. Performance comparison of extracted keywords from different</head><p>Accordingly, we tried high coefficient values on NW DS (ek) and relatively low coefficient values on NW RS (ek) and NW US (ek). However, this scheme produce better results than other schemes as shown in Figure <ref type="figure" target="#fig_4">2</ref>.</p><p>We also added tag frequency information, denoting how many times a tag was This tag frequency rate is calculated as follows.</p><p>( ) ≤ 1, annotated for a denote the set of all tags and the set of all documents</p><p>The above four factors are linearly combined with appropriate coefficients. We nearly optimal of extracted keywords from</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>RS or US). each keyword set on the training dataset comparison of extracted keywords from different</head><p>) and relatively low However, this scheme does not In Figure <ref type="figure" target="#fig_4">2</ref>, Uniform denotes the case of assigning an equal coefficient (0.3) to each keyword set and DS (RS or US) denotes the case of assigning 0.45 to and 0.25 to the other keyword sets. cases. On the contrary to our expectation, the weighting scheme assigning high coefficient value to US</p><p>The reason for this performance of the candidates extracted data columns. Such keywords are In Table <ref type="table" target="#tab_3">3</ref>, it is observed that even the same keyword from different accuracy ratio values. For example, the keyword portal from description has much However, the accuracy ratio than the averages.</p><p>After several trials which has shown fine results on the training dataset NW DS (ek)ⅹ.2 +</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>6</head><p>. Performance comparison among different weighting schemes.</p><p>In Figure <ref type="figure" target="#fig_4">2</ref>, Uniform denotes the case of assigning an equal coefficient (0.3) to each keyword set and DS (RS or US) denotes the case of assigning 0.45 to DS (RS other keyword sets. TFR(ek) was assigned 0.1 or 0.05 in the above On the contrary to our expectation, the weighting scheme assigning high US showed the best performance. e reason for this phenomenon is not clear but one possible clue is that of the candidates extracted from DS varies much according to extracted data columns. Such keywords are illustrated in Table <ref type="table" target="#tab_3">3</ref> In Table <ref type="table" target="#tab_3">3</ref>, it is observed that even the same keyword from DS could have extremely different accuracy ratio values. For example, the keyword portal from has much higher AR(k) value than the average, i.e., about 0.05964. However, the accuracy ratios of the same keyword from url or description After several trials, we applied the following formula for the recommendation, results on the training dataset.</p><p>.2 + NW RS (ek)ⅹ.35 + NW US (ek)ⅹ.4 + TFR(ek)ⅹ.05.</p><p>. Performance comparison among different weighting schemes.</p><p>In Figure <ref type="figure" target="#fig_4">2</ref>, Uniform denotes the case of assigning an equal coefficient (0.3) to each RS or US) in the above On the contrary to our expectation, the weighting scheme assigning high on is not clear but one possible clue is that varies much according to extracted of the same keywords extracted Accuracy values higher than the average are could have extremely different accuracy ratio values. For example, the keyword portal from extended about 0.05964. are lower ing formula for the recommendation,</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Experimental Evaluation</head><p>To evaluate the proposed approach, we reserved the postings spanning the latest six months from the given training dataset like the real challenge. Hence, the training period is from January 1995 to June 2008 and the validation period is from July to December of 2008. The numbers of postings, resources, and users during these periods are shown in Tables <ref type="table" target="#tab_121">4 and 5</ref>.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Effectiveness of Candidate Elimination</head><p>In this subsection, we present the effect of our keyword elimination method (Equation (3)). Note that Limit Condition is applied to the candidate keywords whose accuracy ratio is lower than average with some penalizing effect on frequently-occurred keywords. Figures <ref type="figure" target="#fig_52">3 and 4</ref> show the effect of candidate elimination on the Post-Core and Cleaned Dump datasets, respectively. The results are obtained when the number of recommended tags is five. On the both validation datasets (i.e., Post-Core and Cleaned Dump), the proposed elimination method increases precision and F-measure values regardless of the number of recommended tags (from one to ten, although the results are not shown here). In the case of the Cleaned Dump dataset, recall is also improved by our filtering method.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Conclusion</head><p>We applied a simple weighting scheme for combining different information sources and a candidate filtering method for tag recommendation. The proposed filtering method was shown to improve precision and F-measure for the tag recommendation task in all the cases of our experiments. It has also shown to be effective for improving recall in some cases. Future works include finding more optimal scheme for combining multiple information sources. Evolutionary algorithms would be a suitable methodology for this task.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>The role of the end-user in the world wide web (WWW) has undergone a substantial change in recent years from a passive consumer of relatively static web pages to a central content producer. The addition of backchannels from WWW clients to internet servers empowers non-expert users to actively participate in the generation of web content. This has led to a new paradigm of usage, colloquially coined "Web 2.0" <ref type="bibr">[1]</ref>.</p><p>These novel kinds of interactions can be divided into two categories: producing or making accessible of new information (e.g., web logs, forums, wiki wikis, etc.) and enriching already existing contents (e.g., consumer reviews, recommendations, tagging, etc.). One interpretation of the second type of interaction is that it provides means to cope with one of the problems generated by the first type of interaction, namely the massive growth of available content and the increasing difficulty for traditional information retrieval approaches to support efficient access to the contained information. In this sense, the meta-content produced in the second kind of interaction can be construed as mainly serving as a navigation aid in an increasingly complex, but weakly structured online-world.</p><p>In particular, the possibility for users to attach keywords to web resources in order to describe or classify their contents bares an enormous potential for structuring information which facilitates subsequent access by both the original user and other users. This task, commonly referred to as tagging is simple enough not to scare users away, yet the benefit of web resources annotated in such a way is obvious enough to keep the motivation to supply tags high. The process of attaching tags to web resource must therefore show a fine balance between simplicity and quality. Both of these properties could be greatly improved if an automatic system could support a user by recommending tags for a given resource.</p><p>In this paper, we describe the ARKTiS system, developed to recommend tags to a user for two specific types of web resources. The system was developed as a contribution to an international challenge as part of the 2009 European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML PKDD 2009).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Task and Data</head><p>The goal of this challenge is the implementation of a system that can automatically generate tag recommendations<ref type="foot" target="#foot_32">1</ref> for a given resource. Here, a resource is either a bookmark of a web page or a BibT E X entry for different kinds of documents. Tags typically are short English words although they may also be artificially created terms, generated e.g. by concatenating words ("indexforum") or by using abbreviations ("langfr", for "language french"). The number of tags a system may generate is restricted to a maximum of five.</p><p>To prepare for the challenge, a training set of manually tagged resources was provided. The data set consists of web page bookmarks and BibT E X entries taken from the BibSonomy project <ref type="foot" target="#foot_33">2</ref> . Each entry has a unique id and was tagged by at least two users. Thus a point in the data set can be viewed as a triple &lt;resource-id, user-id, tags&gt;.</p><p>For each resource, the corpus contains meta-data describing the resource with a number of different fields. These fields are different for BibT E X entries and for bookmark entries. For instance, the meta-data for BibT E X entries contain fields describing the title of an article, the authors, the year of publication, or the number of pages. For bookmarks, one field gives a short description of the resource while another one contains the web pages' URL. A full list of all available fields can be found on the homepage of the challenge.</p><p>In total, the training corpus contains 41,268 entries for bookmarks and 22,852 BibT E X entries, annotated with 1,3276 unique tags by 1,185 different users.</p><p>The data of the actual challenge (the eval set), was provided 48 hours before the submission deadline. This set consists of unseen data that each system had to tag and all results that are presented and analyzed in section 5 were achieved on this set. The evaluation data set contains 43,002 entries in total, with 26,104 BibT E X-and 16,898 bookmark-entries -circa two thirds of the amount of the training data.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Approach</head><p>One key observation for our participation in the ECML PKDD challenge was that time played a central role in two different senses. First, the task required each system to suggest tags for quite a large number of resources in a rather short period of time: 43,002 entries in only 48 hours, including retrieval and parsing of the test data as well as formatting and uploading of the final result data. Second, since the challenge had a fixed deadline, the time to develop a running system faced a naturally limit with the release of the 48 hour evaluation period.</p><p>Both points had a direct influence on the conceptualization and realization of our system ARKTiS, in that a number of more sophisticated ideas had to be sacrificed. As a result, ARKTiS can be seen as an exercise in software engineering rather than thorough science. The final system implements straight-forward strategies with a focus on robustness and processing speed.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Motivation</head><p>As outlined above, the training corpus contains 13276 different tags for over 41268 data points, a proportion that indicates a potential data sparseness problem for classical machine learning approaches. We therefore opted for an algorithmic approach instead, based on heuristic considerations. Since the desired system output is (English) words, we can distinguish two potential sources for output: from within the resource itself (internal words) or from outside material (external words). This distinction is in so far blurred, as the system input actually consists not of the resources themselves, but rather of meta-data. In so far, even tags taken from the resources themselves could be argued to be external. We take the view that words stemming from meta-data or the resources referred to by the meta-data are considered internal.</p><p>Since we could not hope to implement a competitive system, we were mainly interested in how useful such a distinction would be in terms of recommending tags. Although we concentrated on internal methods as described below, we explored using document similarity measures in the BibT E X module to re-use tags that were manually assigned to documents similar to the current system input. Also, some of our implemented techniques, such as translating German words from the original resource to English, can be considered borderline between internal and external.</p><p>For the internal approaches, we looked at the task of tagging a resource as an analogy to automatic text summarization, somewhat taken to an extreme where a "summary" consists only of five words. In extractive summarization, summaries of documents are generated by identifying those sentences in a document which when concatenated serve as surrogate for the original document. In that spirit, tagging becomes the task of identifying those words from a resource that together describe the "aboutness" of the resource.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Related Work</head><p>A number of researchers in recent years have engaged in the task of developing an automatic tagging system. <ref type="bibr">[2]</ref> use a natural language resources to suggest tags for BibT E X entries and web pages. They include conceptual information external resources, such as WordNet <ref type="bibr">[3]</ref>, to create the notion of a "concept space". On top of this notion, they exploit the textual content that can be associated with bookmarks, documents and users and generate models within the concept space or the tag space derived from the training data.</p><p>[4] model the problem of automated tag suggestion as a multi-label text classification problem with tags as categories.</p><p>In <ref type="bibr">[5]</ref>, the TagAssist system focuses on the task of the generation of tags for web-log entries. They access tags of similar documents in a similar spirit to our own method described in section 4.1.</p><p>In addition to these concrete systems, we find automatic tagging to bare some similarities with research in automatic extractive summarization. In both task, the identification of salient portions of a resource's text is a central consideration. For tagging which reduces extraction single words, we call such a method an internal approach. (s. section 3.1).</p><p>For instance, in <ref type="bibr">[6]</ref>, the author conducted a wide range of tests to find predictive features for relevant sentences. Despite relying on manual experiments, the general results from this early research were later confirmed by machine learning approaches, e.g., <ref type="bibr">[7]</ref>. For the bookmark modules, both the title and the first sentence heuristic (see 4.2) were inspired by these findings.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Taggers</head><p>As hinted by the data set, the task can be viewed as two different sub-tasks, the tagging of BibT E X entries and the tagging of bookmarked web pages. Consequently, the ARKTiS system consists of two independent modules which are both instances of a common framework architecture, depicted in Figure <ref type="figure" target="#fig_0">1</ref>. The modules can be run in two distinct processes.</p><p>Efficient processing of the input data is an important requirement for this challenge. In a sequential architecture, processing 43002 data points in 48 hours would leave a tagging system about 4 seconds per data point on average. Given that the data points contain only metadata and that the actual documents, if needed, have to be retrieved through the internet and parsed, a time span of 4 seconds poses quite a strong limitation on the complexity of the performed computations. Running multiple instances of taggers concurrently relaxes this limitation. In our current setup, both modules for BibT E X and bookmark tagging internally run ten tagging threads in parallel which increases the maximum average processing time to 80 seconds per data point. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">The BibT E XTagger</head><p>The tagging system responsible for the BibT E X entries uses a combination of internal and external techniques. A thorough investigation of the provided training material showed that most of the entries (95.4%) do not contain a valid link to the actual PDF document. This is unfortunate, as it limits internal approaches which draw tags from the contents of the document in question.</p><p>Internal approach To compensate the cases in which the PDF document is unavailable, we use the remaining information from the meta-data, namely the title of the document, its description and its abstract. The employed approach analyzes these fields and extracts tags directly out of their textual information. Before that, we lowercase all words in the text of each field and remove all punctuation and symbols. After that, we apply the POS tagging system described in <ref type="bibr">[8]</ref> to extract content-words -nouns, adjectives and verbs -out of the text. The sequential processing of the text is shown below: After removing stop-words, the remaining words are directly used as tags, giving preference to tags stemming from the title field over those from description field over those from the abstract field. Only the first five tags are returned after filtering out duplicates.</p><p>External approach For the remaining entries -where the source documents were available -we use a corpus-based approach inspired by standard information retrieval techniques. The idea here is that if a new document is similar to a document from the training corpus, we may re-use the tags that have been added manually to the training document.</p><p>Hence, this part of the tagger first ranks all documents from the training data by similarity to the current document. A second step then takes tags from the documents in rank order and returns the first five of them, discarding duplicates.</p><p>To do so, all documents in the corpus are first transformed from PDF format to plain text by the PDFBox toolkit<ref type="foot" target="#foot_34">3</ref> . After that, the whole text is segmented into sentences using punctuation information (.!?;:\n) and then pre-processed in the same way as described in the internal approach. After removing noncontent words, we calculate tf.idf values for each word in the document, resulting in the following mapping: &lt;list of tags&gt; → &lt;vector of TF/IDF-values&gt; A tf.idf value is a value that calculates the relative importance of the word w i for the current document j, in relation to a set of documents D (see equation 1, where n i,j is the number of occurrence of the word i in document j).</p><formula xml:id="formula_86">tf idf i,j = n i,j k n k,j * log |D| |{d ∈ D : w i ∈ d}|<label>(1)</label></formula><p>This procedure is, of course, carried out only once and the resulting mapping is stored offline. In the actual tagging process, we generate the vector of tf.idf values in the same way for the document to tag and compare the resulting vector to with all document vectors in the corpus. We have experimented with two different similarity measures.</p><p>The first variant compares of two documents by the normalized distance between their tf.idf vectors, as shown in equation 2.</p><formula xml:id="formula_87">sim(t 0 , t 1 ) = N −1 i=0 |t 0 [i] − t 1 [i]| N −1 i=0 t 0 [i] + t 1 [i] (2)</formula><p>In addition, we implemented cosine similarity that measures similarity by the cosine of the angle Θ between the two vectors that the two documents describe (equation 3).</p><formula xml:id="formula_88">sim(t 0 , t 1 ) = cos Θ = N −1 i=0 t 0 [i]t 1 [i] N −1 i=0 t 0 [i] N −1 i=0 t 1 [i]<label>(3)</label></formula><p>In our experiments, the normalized distance measure yielded better performance than cosine similarity and consequently we used only the former in the final system.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">The Bookmark Tagger</head><p>As in the case for the BibT E X tagger, the bookmark tagger relies on relatively simple heuristics to determine the keywords to recommend. The input data provides two kinds of information, the URL of the web page to tag and a short description which in some cases is identical to the web page's title string.</p><p>Processing the URL field In our system, the URL is used to fetch the contents of the actual web page, but since the domain name and path may already contain candidate terms, the URL string is also processed itself, in three sequential steps: tokenizing, filtering, and dict/split.</p><p>For the tokenization, the URL is split up at every non-letter non-digit character, such as a forward slash. By matching against a manually crafted blacklist of terms generally considered uninformative, typical artifacts such as "www" or "html" that result from the tokenization process are filtered out. The following examples illustrate these two steps:</p><p>Original URL: http://www.example.com/new-example/de/bibtex.htm Tokenizing:</p><p>http www example com new example de bibtex htm Filtering:</p><p>example new example de bibtex</p><p>Original URL: http://www.coloradoboomerangs.com Tokenizing: http www coloradoboomerangs com Filtering: coloradoboomerangs A dictionary of American English together with a list of the names of all articles in the English Wikipedia<ref type="foot" target="#foot_35">4</ref> of 2007 are used to check if the resulting tokens are actual words. The rationale for incorporating Wikipedia is that it gives additional terms from article titles which often are not found in a dictionary, such as, e.g. technical terms ("bibtex"). If a token cannot be found in either list, we try to split the token up into two sub-tokens which, in case they are both contained in the dictionary, are then used instead of the original tokens. This idea is based on the observation that domain names in particular are sometimes a concatenation of two terms. Applied to the above example, this step generates the following keyword lists:</p><p>Processing the description field The description that is part of the input data is tokenized in the same way. However, no further attempts are made to filter out tokenization artifacts or to split the resulting tokens into sub-parts in case they are not contained in the dictionary. In other words, of the above three steps, only tokenizing is performed on the description of the bookmark.</p><p>Processing the bookmarked web page With the provided URL, the content of the given web page is retrieved at run-time. We do not attempt to detect whether the server returns an actual content page or a specialized message, such as a HTTP 404 Not Found error page. After the HTML content of a web page has been downloaded, three different extraction methods are applied: HTML-meta, title, and first sentence.</p><p>The first method operates on the head section of the document where it locates and parses the &lt;meta&gt; elements "keywords" and "description". The contents of these elements are provided by the author of the HTML document and may contain valuable hints on what the document actually is about. The contents are extracted and then undergo the same tokenizing procedure as described above.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>HTML:</head><p>&lt;html&gt; &lt;head&gt; &lt;meta name=keywords content="example, sample"&gt; &lt;meta name="description" content="A made-up example webpage"&gt; ... &lt;/head&gt; ... &lt;/html&gt; Extracting: example, sample a made-up example webpage Tokenizing: example sample a made up example webpage Also in the head section is the declaration of the title of the document. This is not only intuitively a good source for relevant keywords, but research in the field of automatic text summarization has also shown in the past that headings contain informative content <ref type="bibr">[9]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>HTML:</head><p>&lt;title&gt;Hello, world -again, an example&lt;/title&gt; Extracting: hello, world -again, an example Tokenizing: hello world again an example Another finding from summarization research is that locational cues work well for determining relevant content words. To apply this insight to the task at hand, the third document-based method tokenizes the first sentence of each HTML document in the same manner described above.</p><p>The result of these steps is a set of basic terms. For the final recommendation, two more processing steps are performed, a ranking step and a normalization step.</p><p>Ranking keywords For the ranking, each of the previously extracted keywords is described according to four predefined dimensions: Source, InDict, POS, and Navigational.</p><p>The values for these dimensions are floating point numbers that represent how valuable a keyword is with respect to being among the recommended tags. For instance, analog to the first step described above, the Source dimension may receive one of the following values:</p><p>-URL (= 0.4) -Description (= 1) -HTML-Meta (= 1) -Title (= 0.8) -First sentence (= 0.9)</p><p>The other dimensions describe whether a keyword is found in the English dictionary (and/or list of Wikipedia articles), its part of speech (NN = 1, NNS = 0.9, VBG = 0.8, VERB = 0.5, OTHER = −4) and whether it is found on a blacklist of navigational terms, such as "impress", "home", etc. which was created manually by the authors. As with other heuristics, the idea of using such stigma word lists can also be tracked back to early summarization research, see e.g. <ref type="bibr">[6]</ref>.</p><p>To rank the keywords, a weighted sum of the four values is computed for each keyword. Since a training corpus was available, good practical weights could have been determined with a machine learning approach. Unfortunately, since time was scarce, we had to estimate sensible weights by hand-inspecting the performance of the tagger on selected samples from the training corpus helped in this part of the development.</p><p>Normalization In a final normalization step, a German-English dictionary is used to translate German keywords to English ones and to re-weight those keywords that contain other keywords as sub-strings. In such a case it was speculated that the keyword contained as a sub-string would likely be the more general term and thus its final score was slightly increased.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Results</head><p>In the evaluation, the results of our tagging system ARKTiS had to be compared against the tags that were annotated by a human. The test data were provided 48 hours before the submission deadline. Considering at most five tags per entry, the evaluation uses precision, recall and f-score values as measurements. In the following, we will present our results, that were achieved on this data and compare them against a baseline system. The baseline system predicts the five most common tags from the training data (Figure <ref type="figure" target="#fig_10">3</ref>) to each input entry.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>BibT E X:</head><p>ccp jrr programming genetic algorithms Bookmarks: software indexforum video zzztosort bookmarks</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Fig. 3. Most common tags in the data</head><p>The results of the baseline system are presented in Table <ref type="table" target="#tab_0">1</ref> where we can see a maximum f-score of 0.55%. Comparing this to the results of ARKTiS (Table <ref type="table" target="#tab_1">2</ref>), we can see that our system clearly outperforms the baseline with an f-score of almost 11%. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Conclusion and Future Work</head><p>Our work shows that is is possible to design and implement a basic tag recommender system even with a very limited development time. The two tracks, BibT E X and bookmark tagging, were designed and realized independently but on top of a common, concurrent framework.</p><p>The overall task can be considered challenging, especially if results are evaluated on the basis of recall and precision: our final system scored a rather low 11% f-score. The large number of different gold-standard tags makes this number difficult to interpret; however, it is clear that it leaves room for improvement. The winning entry of the 2009 challenged reached an f-score of 19 percent.</p><p>In a more detailed analysis, we found that the bookmark module outperformed the BibT E X module to some degree. As described above, the two modules employ rather different approaches, thus a next logical step will be to combine the best ideas from both modules.</p><p>The biggest drawback for ARKTiS as described in this paper was the fact that we entered the ECML PKDD challenge at a late point. As a consequence, a number of interesting and more sophisticated ideas had to be left out of the system purely due to the lack of implementation time. For instance, the use of tf.idfscores in the BibT E X module is very limited, as is the use of content terms beyond the first sentence in the bookmark module.</p><p>At the same time, the ARKTiS system has proven its robustness and will be a good starting point for further research in the area of automatic tagging.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Tag Recommendation using Probabilistic Topic Models</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Ralf Krestel and Peter Fankhauser</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>L3S Research Center Leibniz Universität Hannover, Germany</head><p>Abstract. Tagging systems have become major infrastructures on the Web. They allow users to create tags that annotate and categorize content and share them with other users, very helpful in particular for searching multimedia content. However, as tagging is not constrained by a controlled vocabulary and annotation guidelines, tags tend to be noisy and sparse. Especially new resources annotated by only a few users have often rather idiosyncratic tags that do not reflect a common perspective useful for search. In this paper we introduce an approach based on Latent Dirichlet Allocation (LDA) for recommending tags of resources. Resources annotated by many users and thus equipped with a fairly stable and complete tag set are used to elicit latent topics represented as a mixture of description tokens and tags. Based on this, new resources are mapped to latent topics based on their content in order to recommend the most likely tags from the latent topics. We evaluate recall and precision for the bibsonomy benchmark provided within the ECML PKDD Discovery Challenge 2009.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Tagging systems <ref type="bibr">[1]</ref> like Flickr, Last.fm, Delicious or Bibsonomy have become major infrastructures on the Web. These systems allow users to create and manage tags to annotate and categorize content. In social tagging systems like Delicious the user can not only annotate his own content but also content of others. The service offered by these systems is twofold: They allow users to publish content and to search for content, thus tagging also serves two purposes for the user:</p><p>1. Tags help to organize and manage own content, and 2. Find relevant content shared by other users.</p><p>Tag recommendation can focus on one of the two aspects. Personalized tag recommendation helps individual users to annotate their content in order to manage and retrieve their own resources. Collective tag recommendation aims at making resources more visible to other users by recommending tags that facilitate browsing and search.</p><p>However, since tags are not restricted to a certain vocabulary, users can pick any tags they like to describe resources. Thus, these tags can be inconsistent and idiosyncratic, both due to users' personal terminology as well as due to the different purposes tags fulfill <ref type="bibr">[2]</ref>. This reduces the usefulness of tags in particular for resources annotated by only a few users (aka cold start problem in tagging), whereas for popular resources collaborative tagging typically saturates at some point, i.e., the rate of new descriptive tags quickly decreases with the number of users annotating a resource <ref type="bibr">[3]</ref>.</p><p>The main goal of the approach presented in this paper is to overcome the cold start problem for tagging new resources. To this end, we use Latent Dirichlet Allocation (LDA) to elicit latent topics from resources with a fairly stable and complete tag set. The latent topics are represented as a mixture of description tokens like URL, title, and other metadata, and tags, which typically co-occur. Based on this, new resources are mapped to latent topics based on their description in order to recommend the most likely tags from the latent topics.</p><p>The remainder of this paper is organized as follows. In Section 2, we define the problem of tag recommendation more formally, and introduce the approach based on LDA. In Section 3 we present our evaluation results. In Section 4 we discuss related work, and in Section 5 we summarize and outline possible future research directions.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Tag Recommendation</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Problem Definition</head><p>Given a set of resources R, tags T , and users U , the ternary relation X ⊆ R × T × U represents the user specific assignment of tags to resources. T consists of two disjoint sets T tag and T desc . T tag contains all user assigned tags, T desc contains the vocabulary of content and meta information, such as abstract or resource description, which is represented as tag assignment by a special "user". A post b(r i , u j ) for resource r i ∈ R and a user u j ∈ U comprises all tags assigned by u j to r i : b(r i , u j ) = π t σ ri,uj X<ref type="foot" target="#foot_36">1</ref> . The goal of collective tag recommendation is to suggest tags to a user u j for a resource r i based on tag assignments to other resources by other users collected in Y = σ r =ri∨u =uj π r,t X ⊆ R × T .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Latent Dirichlet Allocation</head><p>The general idea of Latent Dirichlet Allocation (LDA) is based on the hypothesis that a person writing a document has certain topics in mind. To write about a topic then means to pick a word with a certain probability from the pool of words of that topic. A whole document can then be represented as a mixture of different topics. When the author of a document is one person, these topics reflect the person's view of a document and her particular vocabulary. In the context of tagging systems where multiple users are annotating resources, the resulting topics reflect a collaborative shared view of the document and the tags of the topics reflect a common vocabulary to describe the document.</p><p>More generally, LDA helps to explain the similarity of data by grouping features of this data into unobserved sets. A mixture of these sets then constitutes the observable data. The method was first introduced by Blei, et. al. <ref type="bibr">[4]</ref> and applied to solve various tasks including topic identification <ref type="bibr">[5]</ref>, entity resolution <ref type="bibr">[6]</ref>, and Web spam classification <ref type="bibr">[7]</ref>.</p><p>The modeling process of LDA can be described as finding a mixture of topics for each resource, i.e., P (z | d), with each topic described by terms following another probability distribution, i.e., P (t | z). This can be formalized as</p><formula xml:id="formula_89">P (t i | d) = N j=1 P (t i |z i = j)P (z i = j | d),<label>(1)</label></formula><p>where P (t i ) is the probability of the ith term for a given document and z i is the latent topic. P (t i |z i = j) is the probability of t i within topic j. P (z i = j) is the probability of picking a term from topic j in the document. These probability distributions are specified by LDA using Dirichlet distributions. The number of latent topics N has to be defined in advance and allows to adjust the degree of specialization of the latent topics. The algorithm has to estimate the parameters of an LDA model from an unlabeled corpus of documents given the two Dirichlet priors and a fixed number of topics. Gibbs sampling <ref type="bibr">[5]</ref> is one possible approach to this end: It iterates multiple times over each tag t, and samples a new topic j for the tag based on the probability P (z i = j|t, z −i ), where z −i represents all topic-word and document-topic assignments except the current assignment z i for tag t, until the LDA model parameters converge.</p><p>Application to Tagging Systems LDA assigns to each document latent topics together with a probability value that each topic contributes to the overall document. For tagging systems the documents are resources r ∈ R, and each resource in addition to its description from T desc is described by tags t ∈ T tag assigned by users u ∈ U . Instead of documents composed of terms, we have resources composed of tags. To build an LDA model we need resources and associated tags previously assigned by users. For each resource r we need some posts b(r, u i ) assigned by users u i , i ∈ {1 . . . n}. Note that for each resource, at least the tag assignments from its description is available. Then we can represent each resource in the system not with its actual tags but with the tags from topics discovered by LDA.</p><p>For a new resource r new with few or no posts, we can expand the latent topic representation of this resource with the top tags of each latent topic. To accomodate the fact of some tags being added by multiple users whereas others are only added by one or two users we can use the probabilities that LDA assigns. As formalized in Equation 1 this is a two level process. Probabilities are assigned not only to the latent topics for a single resource but also to each tag within a latent topic to indicate the probability of this tag being part of that particular topic. We represent each resource r i as the probabilities P (z j |r i ) for each latent topic z j ∈ Z. Every topic z j is represented as the probabilities P (t n |z j ) for each tag t n ∈ T . By combining these two probabilities for each tag for r new , we get a probability value for each tag that can be interpreted similarly as the tag frequency of a resource. Setting a threshold allows to adjust the number of recommended tags and emphasis can be shifted from recall to precision. Imagine a resource with the following tags: "photo", "photography", and "howto". Table <ref type="table" target="#tab_0">1</ref> shows the top terms for two topics related with the assigned tags. The latent topics comprise a broad notion of (digital) photography and the various aspects of tutorial material. Given these topics we can easily extend the current tag set or recommend new tags to users by looking at the latent topics. If LDA assumes that our resource in question belongs to 66% to the "photo"topic and to 33% to the "howto"-topic, these probabilities are multiplied with the individual topic/tag probabilities, and the top five tags recommended are "tutorial", "howto", "images", "photo", and "photography".</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Evaluation</head><p>We used the data provided by the ECML PKDD Discovery Challenge 2009 to evaluate our approach and fine-tune our parameters. For assessing precision, recall, and f-measure we used the supplied evaluation script.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Dataset</head><p>Our dataset consists of the provided training data for the Discovery Challenge. All experiments were performed on the post-core at Level 2, where all tags, users, and resources occur at least in two posts. To measure the performance of our system, we split the training data into a 90% training set and a 10% test set based on posts (called content IDs in the dataset). For each resource, as defined by the hash values, we build up a textual representation. This representation contains all the tags that were assigned by users in the training set to a particular resource. In addition, we add terms extracted from the description of the resource. More precisely, we tokenized different fields describing a bookmark or bibtex entry. An overview of the fields can be seen in Table <ref type="table" target="#tab_1">2</ref>. Afterwards, we removed stopwords and punctuation marks. Using also the description ensures that we have some terms related to a resource even if no other user before tagged it.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Results</head><p>The tag recommendation algorithm is implemented in Java. We used Mallet <ref type="bibr">[8]</ref>, which provides an efficient SparseLDA implementation <ref type="bibr">[9]</ref>, to perform the Latent Dirichlet Allocation with Gibbs sampling. The LDA algorithm takes three input parameters: the number of terms to represent a latent topic, the number of latent topics to represent a document, and the overall number of latent topics to be identified in the given corpus.</p><p>Table <ref type="table" target="#tab_3">3</ref> shows the actual tag distribution for a randomly selected resource (http://jo.irisson.free.fr/bstdatabase/), the top tags recommended by LDA with aggregated probabilities, and all the tags provided by a sample user. As the actual tags indicate, the url is a database/latex related site. The tags recommended by LDA come from six latent topics, comprising latex, databases, academia, references, bibliography, and style. These tags characterize the resource quite well. Table <ref type="table" target="#tab_5">4</ref> compares the f-measure reached for various numbers of latent topics and the baseline which simply recommends the top most frequent tags for each resource (mf)<ref type="foot" target="#foot_37">2</ref> . As can be seen, the best f-measure for LDA is reached between 2500 and 5000 latent topics, but it does not reach the baseline by far. The main reason for this seems to be that the average number of tags per resource is just 10.3 (7.4 distinct tags). This is significantly smaller than the number of (distinct) tokens in a full-text abstract or document, to which LDA has been applied traditionally. Moreover, there are only about 2.8 posts per resource. Thus, there is on the one hand too little co-occurrence evidence for eliciting latent topics, on the other hand there is too little overlap between users on a resource to effectively predict tags via the latent topics of a resource for a new post.</p><p>However, to deal with resources that have only few tags associated it makes sense to combine tag recommendations based on most frequent tags with tag recommendations based on latent topics. With f req(t, r) the frequency of tag t annotated for resource r, one estimate of the probability of tag t given resource r is as follows:</p><formula xml:id="formula_90">P 1 (t | r) = f req(t, r) ti∈r f req(t i , r)<label>(2)</label></formula><p>This estimate can be combined with the estimate P 2 (t | r) via latent topics in Equation 1 by means of a mixture:</p><formula xml:id="formula_91">P (t | r) = λP 1 (t | r) + (1 − λ)P 2 (t | r).<label>(3)</label></formula><p>Table <ref type="table" target="#tab_7">5</ref> shows that this combination achieves consistently better recall and precision than the individual approaches. The largest gain is achieved for the first recommended tag. Similar accuracies are achieved when varying the mixture parameter λ between 0.3 and 0.9, and for a number of latent topics ≥ 1000.  <ref type="table" target="#tab_10">6</ref> compares the results for 5000 latent topics with the results using the most frequent tags, and the combination of the two approaches <ref type="foot" target="#foot_38">3</ref> . Because for the posts in the test set there are only about 0.3 posts per resource in the training set, recommending only the most frequent tags does not recommend any tags for most of the resources. Consequently, recall and precision are significantly lower than for the approach based on latent topics. The combination of the two approaches achieves slightly but consistently better recall and precision.</p><p>Task 2 operates on the post-core at Level 2, where all tags, users, and resources occur at least twice in the training data, which comprises about 750 K tokens for 22389 resources. The test set consists of 778 posts, for which there exist on average 5.8 posts in the training set. for the two individual approaches and their combination <ref type="foot" target="#foot_39">4</ref> . As is to be expected, recall and precision are much better than for Task 1, because there is more knowledge available about the tagging practices of users. Like in our internal tests tag recommendation based on most frequent tags outperforms the approach based on LDA, and the combination outperforms the individual approaches. Table <ref type="table" target="#tab_19">8</ref> shows the results when only tags are used to elicit latent topics. Recall and precision are consistently lower. Thus taking into account the content of resources leads to more effective latent topics for tag recommendation. However, this does not hold for tag recommendation based on most frequent tags. Recommending the most frequent content terms or tags consistently leads to lower precision and recall.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Related Work</head><p>Tag recommendation has received considerable interest in recent years. Most work has focused on personalized tag recommendation, suggesting tags to the user bookmarking a new resource: This is often done using collaborative filtering, taking into account similarities between users, resources, and tags. <ref type="bibr">[10]</ref> introduces an approach to recommend tags for weblogs, based on similar weblogs tagged by the same user. Chirita et al. <ref type="bibr">[11]</ref> realize this idea for the personal desktop, recommending tags for web resources by retrieving and ranking tags from similar documents on the desktop. <ref type="bibr">[12]</ref> aims at recommending a few descriptive tags to users by rewarding co-occuring tags that have been assigned by the same user, penalizing co-occuring tags that have been assigned by different users, and boosting tags with high descriptiveness (TFIDF).</p><p>Sigurbjörnsson and van Zwol <ref type="bibr">[13]</ref> also look at co-occurence of tags to recommend tags based on a user defined set of tags. The co-occuring tags are then ranked and promoted based on e.g. descriptiveness. Jaeschke et al. <ref type="bibr">[14]</ref> compare two variants of collaborative filtering and Folkrank, a graph based algorithm for personalized tag recommendation. For collaborative filtering, once the similarity between users on tags, and once the similarity between users on resources is used for recommendation. Folkrank uses random walk techniques on the userresource-tag (URT) graph based on the idea that popular users, resources, and tags can reinforce each other. These algorithms take co-occurrence of tags into account only indirectly, via the URT graph. Symeonidis et al. <ref type="bibr">[15]</ref> employ dimensionality reduction to personalized tag recommendation. Whereas <ref type="bibr">[14]</ref> operate on the URT graph directly, <ref type="bibr">[15]</ref> use generalized techniques of SVD (Singular Value Decomposition) for n-dimensional tensors. The 3 dimensional tensor corresponding to the URT graph is unfolded into 3 matrices, which are reduced by means of SVD individually, and combined again to arrive at a more dense URT tensor approximating the original graph. Tag recommendation then suggests tags to users, if their weight is above some threshold.</p><p>An interactive approach is presented in <ref type="bibr">[16]</ref>. After the user enters a tag for a new resource, the algorithm recommends tags based on co-occurence of tags for resources which the user or others used together in the past. After each tag the user assigns or selects, the set is narrowed down to make the tags more specific. In <ref type="bibr">[17]</ref>, Shepitsen et al. propose a recommendation system based on hierarchical clustering of the tag space. The recommended resources are identified using user profiles and tag clusters to personalize the recommendation results. Note that they use tag clusters to recommened resources whereas we use LDA topics, which can be considered clusters, to recommend tags. <ref type="bibr">[3]</ref> introduce an approach to tag recommendation using association rules. Resources are regarded as baskets consisting of tags, from which association rules of the form T 1 → T 2 are mined. On this basis tags in T 2 are recommended whenever the resource contains all tags in T 1 . A comparison of this approach with the approach presented in this paper can be found in <ref type="bibr">[18]</ref>.</p><p>When content of resources is available, tag recommendation can also be approached as a classification problem, predicting tags from content. A recent approach in this direction is presented in <ref type="bibr">[19]</ref>. They cluster the document-termtag matrix after an approximate dimensionality reduction, and obtain a ranked membership of tags to clusters. Tags for new resources are recommended by classifying the resources into clusters, and ranking the cluster tags accordingly.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Conclusions and Future Work</head><p>In this paper we have presented and evaluated the use of Latent Dirichlet Allocation for collective tag recommendation. Using selected features from the content of resources, tags, and users, we elicit latent topics that comprise typically cooccuring tags and users. On this basis we can recommend tags for new users and resources by mapping them to the latent topics and choosing the most likely tags from the topics. The approach complements simple tag recommendation based on most frequent tags especially for new resources with only few posts. Consequently, combining tag recommendations based on latent topics with tag recommendations based on most frequent tags outperforms the individual approaches.</p><p>For future work we want to investigate approaches that take into account individual tagging practices for personalized tag recommendation.</p><p>Regarding data sets, we also want to experiment with datasets from different domains, to check whether photo, video, or music tagging sites show different system behavior influencing our algorithms. Another interesting direction we want to follow is to apply LDA not only for tag recommendation but to employ it in the context of recommending resources.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Folksonomy is a way to categorize Web resources via utilizing the "wisdom" of web users, nowadays it is existing in many web applications such as Delicious<ref type="foot" target="#foot_41">3</ref> , Filckr<ref type="foot" target="#foot_42">4</ref> , Bibsonomy <ref type="foot" target="#foot_43">5</ref> . One user could create and share her knowledge during the tagging on resources that are interesting to her. Web resources come in many forms, for example, one resource could be a Web pages, a published paper, or a book. To tag a resource with appropriate words is not so easy and might cost lots of time. Thus a tag recommendation system is needed for easing the timeconsuming step. Typically a recommendation system would suggest 5 or 10 tags to the user for a given resource. Those suggested tags would help one user to think about eligible words and to realize the interesting aspects concerned by others. To solve the problems, ECML PKDD holds the second round discovery challenge<ref type="foot" target="#foot_44">6</ref> of tag recommendation. This paper presents a probabilistic ranking approach submitted to the challenge.</p><p>Given a resource, users choose tags by different aspects of the resource and their specific interests. To pick up a tag from the entire tag set and assign it to the resource could be formulated as following process: given a resource and a user, ranking the tags by their relevance to the resource and user. Here relevance denotes the 'value' of how likely the user would label this tag on this resource.</p><p>We suppose a tag recommendation system works best while recommending tags are sorted by the relevance and then suggested to the user.</p><p>In this paper, the datasets provided by Bibsonomy is a set of post. Each post denotes a triple {user, resource, a set of tag}. A resource type could be bookmark or bibtex, where bookmark is Web page and bibtex is publication. Both bookmark and bibtex resources contain many fields: URL, description, etc. The textural information in the fields could be merged as a pseudo document.</p><p>A natural way of choosing tags is to select words from the pseudo document of given resource. A TF-like maximum likelihood method could reach the goal. The important problem is that maximum likelihood model could not generate tags which are meaningful but not existing in the document. To incorporate previously popular tags and tags preferred by a user, a tag recommendation model could be formulate into language model smoothed via Jelinek-Mercer method as described in Section 3.2. However, the language modeling approach could not learn the word-tag relateness which reflects how other users choose tags for those words in the document. Since the textural information existing in a post could be considered as a parallel corpus -{words in document, tags}, we propose to use the statistical machine translation approach to learn the translation probability from words to tags.</p><p>Finally, we propose a candidate set based tag recommendation algorithm which generates candidate tags from the textual fields of a resource using maximum likelihood and statistical machine translation model. The effectiveness of our approach is validated on the bookmark and bibtex tagging test datasets provided by Bibsonomy. While textural content of a bookmark resource is inadequate, we utilize the tags used within same Domain to extend the candidate set. We also found simple co-occurrence based translation probability estimation performs as good as IBM Model 1 <ref type="bibr">[6]</ref> which uses the EM algorithm to learn the translation probability. An advantage of co-occurrence based approach is its convenience for handling with new training data, since training the model is just counting the co-occurrence of words and tags. However, EM-based approach needs to re-train translation model though iterations which might be time consuming for large scale dataset.</p><p>The rest of this paper is organized as follows. In Section 2 the related work is surveyed. In Section 3 our content based tag recommendation models are presented, and the recommendation algorithm is described in Section 4. In Section 5 we descrbe the data format and preprocessing step, and experimental results are reported in Section 6. Finally in Section 7 we conclude this paper and give out some possible future research issues.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Related Work</head><p>Most of existing tag recommendation approaches are based on the textual information of the resource and previous interests of users. Up to now, the information retrieval, data mining and natural language processing techniques have been used for solving the tag recommendation problem.</p><p>Heymann et al. <ref type="bibr">[1]</ref> use one of the largest crawls from the social bookmarking system Delicious and presents studies of the factors which could impact the performance of tag prediction. The predictability of tags is measured by some method such as entropy based metric. The tag-based association rule is proposed to assist tag predictions. The method of learning the word-tag relateness via association rule needs to tune the confidence and support to find meaningful rules, but we transfer it into the translation probability which could get the converged solution without tuning.</p><p>Tatu et al <ref type="bibr">[2]</ref> uses document and user models derived from the textual content associated with URLs and publications by social bookmarking tool users. The natural language processing techniques are used to extract the concept(Part of Speech, etc.) from the textual information. WordNet<ref type="foot" target="#foot_45">7</ref> are used to stem the concepts and link synonyms. The difference between our work and theirs is that they expand the concept via WordNet, but do not have the word-to-tag translation probability such as from 'eclipse' to 'java'.</p><p>Lipczak <ref type="bibr">[3]</ref> focus on the folksomomies towards individual users, and proposed a three step tag recommendation system which conducts the Personmony based filtering using previously used tags of users after the extraction and retrieving of tags. The recommendation approach in <ref type="bibr">[3]</ref> is similar with our work, but the scores of candidate tags are computed differently. They use the multiply strategy for different factors, but we conduct a weighted sums in which the weight could be set to prefer different components. Besides, we use the statistical machine translation approach to learn the word-tag relateness which is different from model proposed in <ref type="bibr">[3]</ref>.</p><p>Language modeling approach <ref type="bibr">[4]</ref> has been applied in Information Retrieval with lots of smoothing strategies <ref type="bibr">[5]</ref>. The statistical machine translation approaches <ref type="bibr">[6]</ref> shows its theoretical soundness and effectiveness in translation, and Berger et al <ref type="bibr">[7]</ref> and Xue et al <ref type="bibr">[8]</ref> incorporate the statistical translation approaches into information retrieval and automatic question answering fields. The theoretical soundness and effectiveness make it stable to adopt the language modeling and statistical machine translation approach into tag recommendation. The statistical machine translation approach also naturally solve the problem of learning the word-tag relateness of sharing the common tagging knowledge among users.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Content Based Tag Recommendation Models</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Problem Definition</head><p>In this paper, a tag set is denoted as t = {t i } Q i=1 where t i is a single word or term and Q is the number of tags in t.</p><p>The tag recommendation task is to suggest a tag set t for a user U k while given a bookmark/publication resource R j which might be a web page, a book or paper etc. The resource R j contains several fields such as URL, title, description and we denote the resource content as a pseudo document D j . Suppose the recommendation system is required to suggest N tags, it is to find N tags {t i } N i=1 from the entire tag sets with the biggest probability p(t i |U k , D j ).</p><p>For solving the task, a training set S = {S i } K i=1 is given, where S i specifies a triple {t i , U i , D i }. The t i is a tag set, U i ∈ U = {U 1 , ..., U M } is a user and D i ∈ D = {D 1 , ..., D N } is a resource . Then we can learn a tag recommendation model M from S.</p><p>At the testing stage, a testing set T = {T j } P j=1 where T j = {U j , D j } is given. The model M is asked to suggest tag set t j for each T j . After that a groudtruth tag sets G = {g j } P j=1 is used to judge the recommendations {t j } P j=1 , and the performance is get via some evaluation measures such as Precision, Recall and F-measure.</p><p>For a specific user U k , she would have her preference in choosing a word t i as a tag, and if we have this user's information in the training set S, we can formulate this preference as</p><formula xml:id="formula_92">P (t i |U k ) = c(ti;U k ) |U k |</formula><p>where c(t i ; U k ) is frequency of t i be used by user U k , and |U k | is total frequency of all tags used by U k .</p><p>We define the tag generating probability a tag t i for a given user and document tuple {U k , D j } as:</p><formula xml:id="formula_93">P (t i |D j , U k ) = (1 − β)P (t i |D j ) + βP (t i |U k )<label>(1)</label></formula><p>Where β is a trade-off parameter between the resource content and user. Following we will introduce language model and statistical machine translation approaches for estimating P (t i |D j ), and then we will combine them into our final model.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Language Modeling Approach</head><p>A natural and simple way to estimate P (t i |D j ) is to use the maximum likelihood approach as:</p><formula xml:id="formula_94">P ml (t i |D j ) = c(t i ; D j ) |D j |<label>(2)</label></formula><p>Where c(t i ; D j ) is occurrence of t i in D j , and |D j | is document length of D j . The shortcoming of the maximum likelihood estimation is that it could not generate tag which does not exist in D j , thus we introduce language model smoothed via Jelinek-Mercer method <ref type="bibr">[5]</ref> as:</p><formula xml:id="formula_95">P lm (t i |D j ) = (1 − λ)P ml (t i |D j ) + λP ml (t i |C)<label>(3)</label></formula><p>Where λ is the smoothing parameter, and C corresponds to the entire corpus. Actually the smoothing term P (t i |C) could be formulated as the probability of the word t i be used as a tag. We define P (t i |C) as c(ti)</p><p>#tags where #tags is the total number of tags in the training set S. The language modeling approach (3) could be considered as the incorporation of words in the document and previously popular tags of all users.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Statistical Machine Translation Approach</head><p>However, the language modeling approach has not considered word-tag relateness which would be important for tag recommendation. For solving the problem, we further introduce the Statistical Machine Translation(SMT) approach <ref type="bibr">[6]</ref> [7] <ref type="bibr">[8]</ref> for estimating the probability P (t i |D j ):</p><formula xml:id="formula_96">P smt (t i |D j ) = |D j | |D j | + 1 P tr (t i |D j ) + 1 |D j | + 1 P (t i |null)<label>(4)</label></formula><p>Where P (t i |null) could be regarded as the background smoothing model P (t i |C), and a more detailed comparison them could be found in <ref type="bibr">[8]</ref>. P tr (t i |D j ) is the translation probability from D j to ti as following:</p><formula xml:id="formula_97">P tr (t i |D j ) = w∈Dj P tr (t i |w)P ml (w|D)<label>(5)</label></formula><p>To learn the word-word transition probability P tr (t i |w), the EM algorithm could be used. The detail of EM algorithm of learning the word-tag relateness P (t i |w) in Statistical Machine Translation(SMT) Model is described in <ref type="bibr">[6]</ref>. In the training set S = {S j } K j=1 , the parallel corpus of tag and document as S j = {t j , D j } is utilized, and the EM step for learning P (t i |w) can be formulated as: E-Step:</p><formula xml:id="formula_98">P 1 tr (t i |w) = δ −1 w K j=1 c(t i , w; t j , D j )<label>(6)</label></formula><p>M-Step:</p><formula xml:id="formula_99">c(t i , w; t j , D j ) = P (t i |w) P (t i |w 1 ) + ... + P (t i |w o ) #(t i , t j )#(w, D j )<label>(7)</label></formula><p>In Equation (6) δ −1 w = ti K j=1 c(t i , w; t j , D j ) is the normalization factor. In Equation ( <ref type="formula" target="#formula_9">7</ref>) {w 1 , ..., w o } is words contained in D j , #(t i , t j ) and #(w, D j ) is the number of t i in t j and number of w in D j . The convergency of this EM algorithm is proved in <ref type="bibr">[6]</ref>.</p><p>In this paper, we also find that the co-occurrence based translation probability could be helpful in tag recommendation, and we denote it as:</p><formula xml:id="formula_100">P 2 tr (t i |w) = K j=1 #(t i ; t j ) • #(w; D j ) K j=1 #(w; t j , D j )<label>(8)</label></formula><p>Where #(t i ; t j ) denotes the number of tag t i exists in t j and the same to #(w; D j ). This model could be regarded as a simple approximation of the EM based translation model, and it is also effective. Note that the EM based translation probability is denoted as P 1 tr (t i |w) whereas the co-occurrence based translation probability is denoted as P 2 tr (t i |w) hereafter.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4">Final Model</head><p>Now we combine above methods together to get our final model:</p><formula xml:id="formula_101">P f inal (t i |D j , U k ) =λP (t i |C) + βP (t i |U k ) + αP ml (t i |D j ) + γ w P tr (t i |w)P ml (w|D)<label>(9)</label></formula><p>Where λ + β + α + γ = 1 and P tr could be P 1 tr or P 2 tr . Tuning these four parameters is not easy, and thus we split both Cleaned Dump and Post Core dataset into a training set and a validation set respectively, train the model on the training set and set parameters empirically several times for choosing one with better performance on the validation set. We do not illustrate the detail due to space restriction, and in the experiments we found the performance is relatively well while λ = 0.15, β = 0.1, α = 0.05, γ = 0.7. We use these parameters with Cleaned Dump dataset as our final training set for the challenge.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Candidate Set based Tag Recommendation Algorithm</head><p>Since the task of tag recommendation is to suggest tags for given document and user, it is different from the task of Information Retrieval <ref type="bibr">[7]</ref> or Question Answering <ref type="bibr">[8]</ref> where the query/question is given for finding the relevant documents/answers.</p><p>Given a document D j and user U k , we firstly find a recommendation tag candidate set CS from the words in D j , and we also add the top L related words by P tr (t|w) for every word w in D j . Then we compute the P (t i |D j , U k ) for each tag t i ∈ CS. Finally we sort the tags descending according to P (t i |D j , U k ), and return the top N tags as required by the application system. The L is set to be 20 and N is set to 5 in the experiments. In summary, we get this algorithm in Table <ref type="table" target="#tab_0">1</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Data Preparing and Preprocessing</head><p>The dataset we used is download from ECML PKDD Discovery Challenge 2009 <ref type="foot" target="#foot_46">8</ref>which is provided by BibSonomy<ref type="foot" target="#foot_47">9</ref> . There are two datasets: Cleaned Dump and Post Core. The Cleaned Dump contains all public bookmarks and publication posts of BibSonomy until (but not including) 2009-01-01. The Post Core is a subset of the Cleaned Dump, it removes all users, tags, and resources which appear in only one post from Cleaned Dump. Brief statistics of Cleaned Dump and Post Core could be found in Table <ref type="table" target="#tab_1">2</ref>. One tag assignment means one user choose a tag for a resource, and thus one posts could have several tag assignments. The number of posts are shown for bookmark, bibtex, and entire set. The bookmark and bibtex are seperated by '/', and the entire set are illustrated after ':'. There are three tables tas, bookmark, and bibtex in the dataset. The fields of these tables are list in Table <ref type="table" target="#tab_3">3</ref>. For bookmark resource the field 'content type' is 1 and that of bibtex resource is 2. The fields in bold are used to generate the pseudo document D j and the tags t j in the training process. We firstly remove the stop words in the bookmark and bibtex table since they are seldom used as tags and usually meaningless. The stop word list are download from Lextek <ref type="foot" target="#foot_48">10</ref> . Note that we do not remove stop words in the tas file, and the top 5 stop words exist in Post Core and their frequency could be found in Table <ref type="table" target="#tab_5">4</ref>. There are totally 19, 647 and 2, 513 stop word tag assignments in Cleaned Dump and Post Core, corresponds to 1.39% and 0.99% respectively.</p><p>In contrast, the total frequency of stop words in pseudo documents of Cleaned Dump and Post Core are over 588, 907 and 61, 113, which suggest not to consider stop words as tags in most cases. In Table <ref type="table" target="#tab_7">5</ref> we list out the top 10 tags in Cleaned Dump and Post Core. We could see later that the co-occurrence based translation model are likely to generate words which appear more times. The evaluation measure in following experiments are widely used Precision, Recall, and F1-measure. The testing datasets are released by ECML-PKDD challenge in tasks. There are 2 tasks: task 1 and task 2, where task 1 is for content based tag recommendation, and task 2 is for graph based tag recommendation <ref type="foot" target="#foot_49">11</ref> .</p><p>In task 1 the user, resource of a post might not exist before, so the content information of the resource would be critical for tag recommendation. The results indicates that although P 2 tr (Co-occurrence) is more simpler, it is comparable to P 1 tr . In our previous experiment, we also found sometimes the textual information from the bookmark resource are not adequate enough to generate some tags in the post and it needs to be expanded. Instead of using extrinsic resource such as WordNet, we aggregate the tags in the same web site domain for bookmark resource, and use them to expand the recommendations. The reason we don't expand the term in bibtex is because resources in bibtex are publication and the web site provide less information about tags. Also, trying other tag expansion methods would be our future work. We formulate this expansion as P (t i |Site), and the recommendation model for bookmark would become:</p><formula xml:id="formula_102">P f inal ex (t i |D j , U k ) =λP (t i |C) + βP (t i |U k ) + αP ml (t i |D j ) + γ w P tr (t i |w)P ml (w|D) + θP (t i |Site)<label>(10)</label></formula><p>For illustrate the expansions of different domains, we sample some domains and their top used tags with the probability in Table <ref type="table" target="#tab_10">6</ref>. After the tag expansion via the URL domain, the candidates set CS for the recommendation will have top used tags in the same domain of D j . The performance of <ref type="bibr">(10)</ref> with the expansions on the testing set are shown in Table <ref type="table" target="#tab_117">7 and 8</ref>. The performance are shown for only bookmark, only bibtex, and on entire set. The bookmark and bibtex are seperated by '/', and the entire set are illustrated after ':'. We choose the co-occurrence based model P 2 tr in the competition, and actually the performance in terms of F-measure at 5 is also good when using EM-based model P 1</p><p>tr . The F-measure of EM-based model with the same parameters as Table <ref type="table" target="#tab_11">7</ref> for task 1 and task 2 are shown in Table <ref type="table" target="#tab_20">9</ref>. We can find that the P 2 tr and P 1 tr are comparable once again, on F-measure at 1, the Co-occurrence based model are better, but on F-measure at 5, the EM-based model are better. Next we conduct the experiment on each component of our final model ( <ref type="formula" target="#formula_48">9</ref>), the document maximum likelihood method, language model('LM + User Model'), the EM-based translation model P 1 tr (t i |w), and co-occurrence based translation model P 2 tr (t i |w) are chosen. In the 'LM + User Model' we set the parameters α = 0.5, λ = 0.3, β = 0.2, γ = 0. It could be considered as the language model which incorporates the maximum likelihood, the previously tag probability in the whole corpus, and the user's preference model. The performance on both testing datasets of task 1 and task 2 are illustrated in Figure <ref type="figure" target="#fig_4">2</ref>. The x-axis is the top position from top1 to top5 and the y-axis is the value of F-Measure. We only list out the F1 measure because it reflects both precision and recall.</p><p>Table <ref type="table" target="#tab_20">9</ref>. Performance of ( α = 0.15, λ = 0.05, β = 0.05, γ = 0.5, θ = 0.25 for bookmark, α = 0.15, λ = 0.05, β = 0.1, γ = 0.7 for bibtex with P  From the experimental results we can see the translation based models are better than maximum likelihood method and 'LM + User Model' in task 2. The co-occurrence based model are worst in task 1, and the EM-based model is better than co-occurrence based model on both task. We analyze the results of cooccurrence based model on task 1 and find many recommendations are common used tags, because the co-occurrence based model would prefer to generate those tags occurred more times before. This suggest that if the resource/users have been seen before, thus the co-occurrence based model would perform well, if not, then it is better to choose EM based model. The 'LM + User Model' perform best on task 1, but the performance is still lower than that in Table <ref type="table" target="#tab_11">7</ref>, and also, 'LM + User Model' performs worse than translation models on task 2.</p><p>For comparison between EM-based and co-occurrence based model, we pick out several words w with their top translating words t i in both P 1 tr (t i |w)(EMbased) and P 2 tr (t i |w)(Co-occurrence based). The sampling words could be found in Table <ref type="table" target="#tab_21">10</ref>. We could find that in EM-based translation model, the words are most likely to translate into itself. It indicates that we could consider the EMbased translation model as the combination of the maximum likelihood which only generates the word it self and the co-occurrence based translation model which has higher probability to generate other words as tags. The co-occurrence model are likely to generate those popular tags in the corpus, such as 'tools', 'software', 'social'. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7">Conclusion and Future Work</head><p>In this paper we propose a probabilistic ranking approach for tag recommendation. The textual information from the resources and the parallel textual corpus from previously posts are used to learn the language and statistical translation model. Our hybrid probabilistic approach incorporates both the content based textural model and graph structure existing in posts for sharing the common tagging knowledge among users.</p><p>As our future work, we intent to study how to choose parameters via machine learning approaches to avoid heuristic setting. Further more, increasing the extra information of the resources, for example, using the citations(references) of a publication to augment the information of bookmark resource; using other tag expansion techniques; conducting the natural language understanding of the tag concept as well as studying the evaluation measures for tag recommendation are all possible future research work.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Tag Sources for Recommendation in Collaborative Tagging Systems</head><p>Marek Lipczak, Yeming Hu, Yael Kollet, Evangelos Milios Faculty of Computer Science, Dalhousie University, Halifax, Canada, B3H 1W5 lipczak@cs.dal.ca</p><p>Abstract. Collaborative tagging systems are social data repositories, in which users manage resources using descriptive keywords (tags). An important element of collaborative tagging systems is the tag recommender, which proposes a set of tags to each newly posted resource. In this paper we discuss the potential role of three tag sources: resource content as well as resource and user profiles in the tag recommendation system. Our system compiles a set of resource specific tags, which includes tags related to the title and tags previously used to describe the same resource (resource profile). These tags are checked against user profile tags -a rich, but imprecise source of information about user interests. The result is a set of tags related both to the resource and user. Depending on the character of processed posts this set can be an extension of the common tag recommendation sources, namely resource title and resource profile.</p><p>The system was submitted to ECML PKDD Discovery Challenge 2009 for "content-based" and "graph-based" recommendation tasks, in which it took the first and third place respectively.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>The emergence of social data repositories made a fundamental change in the way information is created, stored and perceived. Instead of a rigid hierarchy of folders, collaborative tagging systems (e.g., BibSonomy<ref type="foot" target="#foot_50">1</ref> , del.icio.us<ref type="foot" target="#foot_51">2</ref> , Flickr<ref type="foot" target="#foot_52">3</ref> , Technorati<ref type="foot" target="#foot_53">4</ref> ) use a flexible folksonomy of tags. The folksonomy is created collaboratively by system users. While adding a resource to the system, users are asked to define a set of tags -keywords which describe it and relate it to other resources gathered in the system. To ease this process, some folksonomy services recommend a set of potentially appropriate tags. Proposing a tag recommendation system was a task of ECML PKDD Discovery Challenge 2009 <ref type="foot" target="#foot_54">5</ref> . This paper presents a tag recommendation system submitted to the challenge.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.1">Definitions</head><p>Collaborative tagging systems allow users (u i ∈ U ) to store resources (r j ∈ R) in the form of posts (p ij ∈ P ). A post is a triple p ij = (u i , r j , T ij ), where T ij = {t k } is a set of tags assigned by the user to the resource. The data structure constructed by the collaborative tagging system (referred to as folksonomy <ref type="bibr">[5]</ref>) is simply a set of posts. However, relations between three basic elements of the post allow us to represent the folksonomy as a tripartite graph of resources, users and tags. Each post can be then understood as a set of edges that form triangles connecting resource, user and tag. Projections of this tripartite graph can be used to examine the relations between folksonomy elements (e.g., two tags can be considered as similar when they are both linked to a large number of common resources, two users are similar when they are linked to the same tags).</p><p>Tag recommendation s is a pair (t, l), where t is a tag and l is a recommendation score, which is supposed to reflect the likelihood of the tag t being chosen by a user as a proper tag. A tag recommendation system returns a set of tag recommendations S. In this paper we use the term tag recommendation set (or simply recommendation) not only to refer to the final set of tags returned to the user, but also to denote the results of intermediate tag recommendation steps. In section 5 we define a set of operations on tag recommendation sets, which are used by our tag recommendation system.</p><p>User profile is a set of tags used by the user prior to the post that is being currently added to the system, P u = {t k :</p><formula xml:id="formula_103">u i = u, r j ∈ R, p ij ∈ P, t k ∈ T ij }.</formula><p>The user profile is usually referred to as personomy <ref type="bibr">[5]</ref>. We use a more general term, because it does not imply that the profile is personal. By analogy we can define a resource profile, which contains all tags that were attached to the resource (e.g., a scientific publication) by all users prior to the current post, P r = {t k : u i ∈ U, r j = r, p ij ∈ P, t k ∈ T ij }. Both user and resource profiles can serve as a simple tag recommendation set. For example, resource profile recommendation S Pr is a set of tags from resource profile of r. Their score is the ratio of posts in which the tag was used to all posts of the resource (Eq. 1). The intuition behind this formula is that tags frequently used to describe a resource are likely to be used again, hence they are good recommendations.</p><formula xml:id="formula_104">l(t k , r) = |{p ij : u i ∈ U, r j = r, t k ∈ T ij }| |{p ij : u i ∈ U, r j = r}|<label>(1)</label></formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.2">Tag recommendation tasks</head><p>The off-line evaluation of a tag recommendation system for challenge purposes is a complex task. Tags added to the resource are highly dependent on the state of the system and previous decisions of the user. It is not possible to create a large, realistic test dataset of posts, hiding at the same time the tags used in these posts. A test dataset which is large enough to objectively measure the quality of a recommendation system must cover a long period of time. If the tags in test data are hidden we lose access to the information about the state of the system, especially newly joined users, which make the dataset not representative. To ease this problem, the organizers of ECML PKDD Discovery Challenge 2009 divided the recommendation task into two subtasks which simulate two complementary recommendation approaches. The first task "content-based recommendation" focuses on the content of a resource that is tagged. In this task we assume that information about the resource and user profile is in most cases not available in the folksonomy. A recommender based on resource content is especially important for new users, which are in the early stage of building their profile. Although, as shown in Section 3.1, the need of creating the recommendation based only on the content is rare, the content based recommender can be a valuable starting point for more complex recommenders that use information gathered in the folksonomy. Such more complex recommenders are evaluated in the second task -"graphbased recommendation". The test set in this task contains only users, resources and tags that were present at least twice in the training data. To obtain this set the organizers extracted k-core of order 2 <ref type="bibr">[2]</ref> of tripartite graph of users, resources and tags created from training data. The test set contained only posts for which user, resource and all tags can be found in the k-core. It is important to notice that the second task neglects the disproportion between the number of unique resources and users. It also greatly simplifies the recommendation task by removing posts with unique tags which are hardest to recommend in real systems. To improve the results for this task the system must follow some unrealistic assumptions. Although this paper describes an entry to the challenge, we aimed to present a general system which can be applied to a real folksonomy based repository of bookmarks or scientific publications. Each modification that was made to match the specific constraints created by the dataset and the second task of the challenge is clearly stated.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Related work</head><p>Most of the tag recommendation systems presented in the literature are graphbased methods. It is a natural choice for folksonomies in which textual content is hard to access. For example, a system by Sigurbjörnsson and van Zwol <ref type="bibr">[9]</ref> uses co-occurrence of tags to propose tags that complement user-defined tags of photographs in Flickr. Jäschke et al. <ref type="bibr">[6]</ref> proposed a graph-based recommendation system for social bookmarking services. The method is based on FolkRank, a modification of PageRank, which is suited for folksonomies. The evaluation on a dense core of folksonomy showed that the FolkRank based recommender outperforms PageRank and collaborative filtering methods.</p><p>Even if a tag recommendation system extracts tags from the resource content, usually it also uses the graph information. An example of a content-based recommender is presented by Lee and Chun <ref type="bibr">[7]</ref>. The system recommends tags retrieved from the content of a blog, using an artificial neural network. The network is trained based on statistical information about word frequencies and lexical information about word semantics extracted from WordNet. Another system de-signed to recommend tags for blog posts is TagAssist <ref type="bibr">[10]</ref>. The recommendation is built on tags previously attached to similar resources. Meaning disambiguation is performed based on co-occurrence of tags in the complete repository.</p><p>Finally, we would like to mention two somewhat similar systems which took the first and second place in the ECML PKDD Discovery Challenge 2008. The winning system was proposed by Tatu et al. <ref type="bibr">[11]</ref>, while the second place was taken by our submission <ref type="bibr">[8]</ref>. Both systems utilize information from resource content and the folksonomy graph. The graph is used to create a set of tags related to the resource and a set of tags related to the user who is adding the resource to the system. The winning system bases these sets on tags gathered in the profile of resource or user. Natural language processing techniques are later used to extend the set of tags related to resource or user (i.e., WordNet based search for words that represent the same concept). Our system bases the resource related tags on the resource title, the set is extended by finding tags that co-occur with the base tags in the system. The user related tags are simply the tags from the user profile. The intersection of both sets creates a set of tags that are related to both resource and user. Our system tries to extend this set by finding more related tags in user profile. Finally, both systems extract tags from resource content and join the content tags with the resource and user related tags to create the final recommendation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">BibSonomy dataset</head><p>All presented experiments and the evaluation of proposed tag recommendation system were performed on a snapshot of BibSonomy <ref type="bibr">[4]</ref>, a collaborative tagging system, which is a repository of website bookmarks and scientific publications (represented by BibTeX entries). The training dataset contained posts entered to the system before January 1, 2009. The test data contained posts entered between January 1, 2009 and June 30, 2009. The snapshot was provided by the organizers of the ECML PKDD Discovery Challenge 2009. The preprocessing steps, applied prior to the release of the dataset, included removing useless tags (e.g., system:unfiled ), changing all letters to lower case and removing nonalphabetical and non-numerical characters from tags. We decided to clean the dataset further by removing sets of posts that were imported from an external source. This preprocessing step involved posts for which one set of tags, defined by user or system, was assigned to a large number of imported resources. An example of such a set consists of 9, 183 posts tagged with tag indexforum by one user. Leaving that tag in the system would result in a biased profile of its author. Unfortunately, this cleaning step could not detect another type of imported posts, for which the system automatically defines tags and timestamps based on the information from an external source. An example of such posts is a set of bookmarks imported from a web browser, for which the collaborative tagging system can use the names of bookmark folders to automatically define tags. The second preprocessing step applied to the released data was separation of bookmark and BibTeX posts. We observed that the vocabulary used for both types of resources is different, even for individual users. Some of the tags (e.g., free) have different meaning when tagging websites or scientific publications. Finally, content based recommendation can be based on different metadata fields in both resource types.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">General characteristics</head><p>According to the statistical information about the dataset presented on the Discovery Challenge website <ref type="foot" target="#foot_55">6</ref> the BibSonomy snapshot matches the usual characteristics of folksonomies, including large disproportion between the number of unique resources and users (Table <ref type="table" target="#tab_0">1</ref>). Among the posts in the BibSonomy snapshot 90% contained unique resources. These resources cannot be found in any other post, hence it is not possible to deduce tag recommendation based on resource profile. At the same time 0.8% of the posts, corresponding to 3,167 posts, were entered by users with no previous posts in the system. Except those posts, every time a post is added, the system is able to use the user profile to recommend tags. Similar proportions can be observed for the CiteULike<ref type="foot" target="#foot_56">7</ref> dataset. The disproportion between unique resources and users is ignored in the test data of "graph-based recommendation" task. All users and resources present in the dataset can be found in the training data at least twice. Despite this fact the differences in statistical characteristics of resource and user profiles should be taken into consideration while proposing a recommendation system for this task. The cumulative frequency distribution of resources shows that both for bookmark and BibTeX entries, even if we remove elements that occurred twice or less, most of the remaining elements still have a very small profile (Fig. <ref type="figure" target="#fig_0">1</ref>). Looking at the same statistic for users we see that a significant fraction of them have over 100 posts in their profiles. Hence user profiles are likely to contain more potentially useful tags. To confirm this hypothesis we ran another experiment in which we simulated the test data of "graph-based recommendation" task and checked what is the precision and recall of basic recommenders that propose tags from resource/user profile sorted by frequency against real tags. To obtain a test set we divided the training data into training posts (entered before September 1, 2008) and test posts (entered later). We pruned them to be sure that all resources, users and tags occurred in the remaining part of the training set at least twice. Although this setting favours resource profiles, their overall recall is still lower than recall of the user profiles (Fig. <ref type="figure" target="#fig_4">2</ref>). The fact that resource profiles are smaller makes them, however, a more precise source of tags. High recall of user profiles was observed by us repeatedly in many experiments. This is the reason why in our work we focused on user profiles, trying to increase the precision of this source of tags, while preserving reasonably high recall.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Tag recommendation sources</head><p>The presented recommendation system is the evolution of the work on the system <ref type="bibr">[8]</ref> submitted to the ECML PKDD Discovery Challenge 2008 <ref type="foot" target="#foot_57">8</ref> . In this section we summarize the results of experiments conducted during the work on the previous version of the system. Their main objective was to evaluate the quality of three basic sources of tags -words from resource title, tags assigned to the resource by other users (resource profile) and tags in user profile.</p><p>Resource title We tested most of the metadata fields looking for potential tags. Among them the resource title appears to be the most robust source of tag recommendations. The title is a natural summarization of web page or scientific publication, which means it plays a similar role as tags. In addition, the title is present on the resource posting page, which means it can possibly suggest tags to the user. It is easy to notice the evidence for this observation in the example posts of User B and User C shown in Table <ref type="table" target="#tab_1">2</ref>. Both of them used the tags prediction and social for "Social tag prediction" paper, which became the only occurrence of these tags in their profiles, unlike tag recommender which was used by them around fifty times, probably to describe the general area of interests.</p><p>The number of words in the title is comparable to the number of tags, hence no additional cleaning steps are needed the achieve fairly high precision comparing to other examined tag sources (around 0.1). The drawback of this source is low recall (around 0.2), which makes the title inappropriate as a stand-alone tag recommender. For bookmark posts the web page URL appears to be another valuable source of tags. Although URL tags are less precise than title tags, their union can increase the recall of recommendation.  <ref type="table" target="#tab_1">2</ref>. Example posts of three users tagging two publications related to the tag recommendation problem (two tags were removed to increase anonymity of posts). Bold tags seem to be suggested by the title. Tags in italics likely represent the concept of tag recommendation problem in users' profiles.</p><p>Resource profile Tags assigned to the resource by other folksonomy users are not a good source of tag recommendations. One of the reasons is the sparsity of data; 90% of resources were added to the system only once. This fact significantly limits the possible recall of this source of tags. The other issue is the personal character of posts and tags, which hurts the precision of retrieved tags. Given the example of two resources about the same concept, we see that users cannot agree on tags describing it: tag recommendation, tag recommender, tagging recommender (Table <ref type="table" target="#tab_1">2</ref>). The variety of tags attached by users creates, however, another application of resource tag sets. Mining relations between tags attached to the same resource can result in a graph of relations between tags. Using a relationship graph the system can identify tags which are also potential recomendations. The graph consists of general relations between tags and can be used independently of the resources, which reduces the negative impact of data sparsity. In our work we use two types of graphs. TagToTag graph is a directed graph which captures the co-occurence of tags. The weight of an edge is analogous to the confidence score (Eq. 2) in association rule mining <ref type="bibr">[1]</ref>, where support({t 1 ∩t 2 }) is the number of co-occurrences of tags t 1 and t 2 and support({t 1 }) is the number of occurrences of tag t 1 . The second graph (TitleToTag) is created specifically for the resource title as the base of the recommendation. Using the same model it captures the relations between words from resource title and its tags.</p><formula xml:id="formula_105">conf idence(t 1 , t 2 ) = support({t 1 ∩ t 2 }) support({t 1 })<label>(2)</label></formula><p>User profile For cognitive simplicity and effcient retrieval, a typical user employs the same limited set of tags to describe resources of the same topic (Table <ref type="table" target="#tab_1">2</ref>). This pattern is the reason for high recall of user tags. On the other hand the user profile is a combination of tags related to many user interests and activities, which makes it a very imprecise source of tags. The most frequent tags from the user profile are likely to be related to the most central interests of the user. In our system we try to utilize the potential of user profile tags to extract user's tags that are related to the interests specific to the posted resource.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Tag recommendation system</head><p>Our tag recommendation system is a composition of six basic tag recommenders (Fig. <ref type="figure" target="#fig_10">3</ref>). The result of each recommender is a tag recommendation set with scores in the range [0, 1]. The recommender makes a decision based on the resource content, resource related tags and user profile tags. However, its design makes it applicable to all posts even if the resource or user profile cannot be found in the system database. In such cases, the corresponding basic recommenders are not active. The following sections and Algorithm 1 give the detailed description of each basic recommender and the data flow in the system.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1">Recommendation based on resource content</head><p>The process starts with the extraction of potential tags from the content of resource. For BibTeX posts the title of publication is used, for bookmarks the  title recommendation is combined with tags extracted from the resource URL. Each word extracted from the title (or URL) is scored based on the usage of this word in previous posts. The score is the ratio of the number of times the word was used in the title (or URL) and as a tag to the total number of occurrences of the word in the title (or URL). Low-frequency words (i.e., words that were used in the title less than 50 times) are assigned an arbitrary score 0.1 which is the estimated probability of using a low-frequency word as a tag. To improve precision, content based recommender tags with score lower than 0.05 are removed from the recommendation set. This step serves also as a language independent stop-words remover. Preliminary experiments indicated that the bookmark title is more precise source of tag recommendation than its URL. This observation should be reflected in the way both tag recommendation sets are merged for bookmark posts. We tested a few rescoring functions, the best results were observed for the leading precision rescorer (Eq. 3), which sets the average precision (based on training data) as the score of first tag l 1 and modifies the scores of following tags l i to preserve the proportion between all tag scores. Based on the tests on training data, the average precision of the title tag with the highest score is 0.2, while for URL it is 0.1.</p><formula xml:id="formula_106">l i = avgP recisionAt1 * l i l 1<label>(3)</label></formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2">Extraction of resource related tags</head><p>The result of title recommender is later used to propose title related tags in TitleToTag recommender. The related tags are extracted for each title word independently. The relation score, multiplied by the score of the word from the title recommender, becomes the score of the tag. This process produces a set of related tags for each title word. These sets are later merged, the scores of tags that can be found in more than one set are summed as they were probabilities of independent probabilistic events (Eq. 4). TagToTag recommender processes tags analogously, however, the input of this recommender is a complete content based tag recommendation set (title and URL for bookmarks). The aim of these recommenders is to produce a large, but likely not precise set of tags related to the resource. The third recommender that is able to produce a similar set is the resource recommender, which returns a set of tags from resource profile. The score of resource tag is the number of its occurrences divided by the number of occurrences of the resource. Although for most real posts this recommender would not return any tags, it plays a significant role in the "graph-based recommendation" task, where the resource of each tested post can be found in the system database at least twice. The scores of the results of three recommenders are summed in a probabilistic way (Eq. 4). This union of tags represents all the tags that are somehow related to the resource, and we refer to them as resource related tags.</p><formula xml:id="formula_107">l merged = 1 − i:ti=t merged (1 − l i )<label>(4)</label></formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3">Recommendation based on user profile</head><p>The user recommender produces a set of tags that were used by the user prior to the current post. Issues related to the construction of user profiles (i.e., import of posts, possible change of user interests) make a simple frequency value not a good score for user profile based recommendation. Tags most likely to be reused are the ones that were steadily assigned to posts while the user profile was built. To capture these tags we counted the number of separate days in which a tag was used by the user. To obtain the tag score we divided the number of days the tag was used by the total number of days in which the user was adding posts to the system. This approach allows a decrease in the importance of tags that were assigned by the user in a short period of time only; however, it only partially solves the problem of imported posts. For some of imported posts the system automatically produces low-quality tags and assigns time stamps copied from an external repository (e.g., importing web browser bookmarks, the system copies the time they were created). The combination of artificial tags and real time-stamps makes these posts very hard to detect. Removing such artificial posts is likely to improve the accuracy of the user profile recommender in a real recommendation system; however, it can have undesired consequences when applied to the challenge datasets. If the user imported posts before both training and test data were collected, it is possible that some of them can be found in both datasets. Hence we should train the system for tags from these posts, because it is possible that they can be found in test data as well. Even if we modify the frequency score the representation of user profile still contains tags related to various user interests. Checking the tags extracted from user profile against resource related tags allows us to extract tags that are particularly important for the processed posts. The intersection of both sets of tags produces tags related both to user as well as resource. The score of a tag is the product of scores from both source sets.</p><p>Finally the results of title recommender, resource recommender and the intersection of resource related tags and user profile are merged. As all three sets are results of independent recommenders, tags must be rescored to ensure that tags from more accurate recommenders will have higher score in the final tag recommendation set. Again the leading precision rescorer was used for the three input tag recommendation sets. The top ten tags of this set create the final recommendation set. The challenge organizers proposed to limit the recommendation set size to five tags, which seems to be a good number to be presented to a user, however, for evaluation purposes it is interesting to observe more tags.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Evaluation</head><p>This section presents the results of the off-line system evaluation based on the available BibSonomy snapshot. The evaluation approach assumed that all and only relevant tags were given by the user. Although this method simplifies the problem, it is robust and objective. The quality metrics were precision and recall, commonly used in recommender system evaluations <ref type="bibr">[3]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.1">Methodology</head><p>To keep the list of correct tags secret during the contest the organizers kept strict division between training and test set. The test data contained posts entered to to BibSonomy between January 1, 2009 and June 30, 2009. Each post of which user, resource and all tags could be found in k-core of order 2 of training data was used as test post for the "graph-based recommendation" task. The remaining posts were used for the "content-based recommendation" task. Comparison of training and test data for both tasks is presented in Table <ref type="table" target="#tab_3">3</ref>.</p><p>As we decided to separate the processing of BibTeX and bookmark posts we present the results for two post types separately. The final recommendation is presented together with the intermediate steps of the system: tags extracted from the resource title (and URL), the most frequent tags from resource profile and user profile and the combination of resource related tags and user profile tags. As each tag from the tag recommendation set can be ranked by its score it is straightforward to present any selected number of recommended tags. The plots (Fig. <ref type="figure" target="#fig_24">6</ref>.1) present consecutive results for the top n tags, where 1 ≤ n ≤ 10. For the "graph-based recommendation" task the tags that could not be found in   </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.2">Results</head><p>As expected, precision and recall of the recommendation results in the "contentbased recommendation" task are mostly driven by the content tags. Low score of user profile recommenders for BibTeX data is likely caused by a large number of posts by users who started to use the system after the training set was built.</p><p>According to the rules set by the organizers the precision score was averaged over all posts in the test set, even if a recommender returned no tags for some of them. Whenever a user profile was available the user based recommender obtained significantly better results than content based recommender only.</p><p>The results for the "graph-based recommendation" task show surprisingly high accuracy of resource profile tags (which was not observed to such a degree on training data). For the test dataset in this task the intersection of resource related tags and user profile has lower precision than resource profile tags. This is an unexpected result, comparing to the previous results on training dataset, where the intersection of resource related tags and user profile had comparable or higher precision and recall to resource profile. Despite this unexpected behaviour the tags from the user profile are able to increase the f1 score by 0.02 for tag recommendation set of size 5. The open question is how representative the results of this dataset are, considering the fact that less than 2% of test posts matched the conditions of this task.</p><p>For both tasks there is a noticeable difference between the results for both types of data. However, it is not clear if it is caused by some fundamental differences between BibTeX and bookmark posts, or the differences between the two particular test datasets used. It is important to notice that the high number of tested posts has no impact on the statistical validity of results. The way the test data was prepared makes it very dependent on the behavior of users in the period of time the data was collected.</p><p>Finally we present the results of the final recommendation for combined Bib-TeX and bookmark posts, which were submitted to the challenge (Table <ref type="table" target="#tab_5">4</ref>). The systems were ranked based on the f1 score (Eq. 5) for the tag recommendation set of size 5. Based on that criterion the presented tag recommendation system took the first place in the "content-based recommendation" task (out of 21 participants) and the third place in the "graph-based recommendation" task (again, out of 21 participants).</p><formula xml:id="formula_108">f 1 = 2 * precision * recall precision + recall<label>(5)</label></formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7">Conclusions and future work</head><p>In creating the presented tag recommendation system we considered the title of a resource as a natural starting point of the recommendation process. We tried to extend the set by tags related to the title as well as tags present in the profiles of resource and user. Our main aim was to extract valuable tags Table <ref type="table" target="#tab_5">4</ref>. The results of the presented tag recommendation system. In the challenge the systems were ranked based on the f1 score for the tag recommendation set of size 5. from user profile which is a very rich but imprecise source of tags. Designing the system we mostly focused on the precision of the recommended tags. To avoid the risk of recommending tags less precise than tags extracted from the title we decided to leave it as the only recommendation whenever the user profile was unavailable. This was a frequent case in "content-based recommendation" task, which gives us hope that the system will be able to achieve even better results for the final "on-line recommendation" task. The system is now connected to BibSonomy and recommends tags to each newly added post in real time. This evaluation setting will give a realistic assessment of system quality.</p><p>In our future work on this project we plan to focus on tagging patterns of individual users which would allow us to tune the recommendation for each specific user. Discovering strong patterns, like user who uses author name and year of publication for each BibTeX post, can greatly increase the accuracy of recommender for this specific user. Another interesting issue is handling of multiword concepts (e.g., is a user going to use two tags "information" "retrieval" or one "information.retrieval"?). Finally, we hope that evaluation settings like "online recommendation" task would allow us to investigate short temporal patterns when a user adds a sequence of posts related to the same problem. Abstract. This work proposes an approach to collaborative tag recommendation based on a machine learning system for probabilistic regression. The goal of the method is to support users of current social network systems by providing a rank of new meaningful tags for a resource. This system provides a ranked tag set and it feeds on different posts depending on the resource for which the recommendation is requested and on the user who requests the recommendation. Different kinds of collaboration among users and resources are introduced. That collaboration adds to the training set additional posts carefully selected according to the interaction among users and/or resources. Furthermore, a selection of post using scoring measures is also proposed including a penalization of oldest post. The performance of these approaches is tested according to F1 but just considering at most the first five tags of the ranking, which is the evaluation measure proposed in ECML PKDD Discovery Challenge 2009. The experiments were carried out over two different kind of data sets of Bibsonomy folksonomy, core and no core, reaching a performance of 26.25% for the former and 6.98% for the latter.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Collaborative Tag Recommendation</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Recently, tag recommendation has been gained popularity as a result of the interest of social networks. This task can be defined as the process of providing promising keywords to the users of a social network in the presence of resources of the network itself. These keywords are called tags and the users can assign them to the resources <ref type="bibr">[12]</ref>. Tagging resources present several advantages: they facilitate other users a later search and browsing, they consolidate the vocabulary of the users, they provide annotated resources and they build user profiles. An option to perform such task could be to provide tags manually to each user, but this time-consuming and tedious task could be avoided using a Tag Recommender System (TRS).</p><p>Folksonomies are examples of large-scale systems that take advantage of a TRS. A Folksonomy <ref type="bibr">[9]</ref> is a set of posts included by a user who has attached This research has been partially supported by the MICINN grants TIN2007-61273 and TIN2008-06247.</p><p>a resource through a tag. Generally, each resource is specific to the user who added it to the system, as Flickr, which shares photos, or BibSonomy, which shares bookmarks and bibtex entries. However, for some types of networks identical resources can be added to the system by different users, as is the case of Del.icio.us which shares bookmarks.</p><p>This paper proposes an approach to collaborative tag recommendation based on a logistic regression learning process. The work starts from the hypothesis that a learning process improves the performance of the recommendation task. It explores several information the learner feeds on. In this sense, the training set depends on each test post and it is specifically built for each of them. In addition, a set of additional posts carefully selected are added to the training set according to the collaboration among users and/or resources.</p><p>The remainder of the paper is structured as follows. Section 2 presents background information about tag recommendation in social networks. Our approach is put in context in Section 3 while the proposed method is provided in Sections 4, 5 and 6. Section 7 describes the performance evaluation metric. The results conducted on public data sets are presented and analyzed in Section 8. Finally, Section 9 draws conclusions and points out some possible challenges to address in the near future.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Related Work</head><p>Different approaches have been proposed to support the users during the tagging process depending on the purpose they were built for. Some of them makes recommendations by analyzing content <ref type="bibr">[1]</ref>, analyzing tag co-occurrences <ref type="bibr">[23]</ref> or studying graph-based approaches <ref type="bibr">[10]</ref>.</p><p>Brooks et al. <ref type="bibr">[4]</ref> analyze the effectiveness of tags for classifying blog entries by measuring the similarity of all articles that share a tag. Jäschke et al. <ref type="bibr">[10]</ref> adapt a user-based collaborative filtering as well as a graph-based recommender built on top of FolkRank. TagAssist <ref type="bibr">[24]</ref> recommends tags of blog posts relying upon tags previously attached to similar resources.</p><p>Lee and Chun <ref type="bibr">[14]</ref> propose an approach based on a hybrid artificial neural network. ConTag <ref type="bibr">[1]</ref> is an approach based on Semantic Web ontologies and Web 2.0 services. CoolRank <ref type="bibr">[2]</ref> utilizes the quantitative value of the tags that users provide for ranking bookmarked web resources. Vojnovic et al. <ref type="bibr">[27]</ref> keep in view collaborative tagging systems where users can attach tags to information objects.</p><p>Basile et al. <ref type="bibr">[3]</ref> propose a smart TRS able to learn from past user interaction as well as from the content of the resources to annotate. Krestel and Chen <ref type="bibr">[13]</ref> raise TRP-Rank (Tag-Resource Pair Rank), an algorithm to measure the quality of tags by manually assessing a seed set and propagating the quality through a graph. Zhao et al. <ref type="bibr">[29]</ref> propose a collaborative filtering approach based on the semantic distance among tags assigned by different users to improve the effectiveness of neighbor selection.</p><p>Katakis et al. <ref type="bibr">[12]</ref> model the automated tag suggestion problem as a multilabel text classification task. If the item to tag exists in the training set, then it suggests the most popular tags for the item. Tatu et al. <ref type="bibr">[25]</ref> use textual content associated with bookmarks to model users and documents.</p><p>Sigurbjornsson et al. <ref type="bibr">[23]</ref> present the results by means of a tag characterization focusing on how users tags photos of Flickr and what information is contained in the tagging.</p><p>Most of these systems require information associated with the content of the resource itself <ref type="bibr">[3]</ref>. Others simply suggest a set of tags as a consequence of a classification rather than providing a ranking of them <ref type="bibr">[12]</ref>. Some of them require a large quantity of supporting data <ref type="bibr">[23]</ref>. The purpose of this work is to avoid these drawbacks using a novel approach which establishes a tag ranking through a machine learning approach based on logistic regression.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Tag Recommender Systems (TRS)</head><p>A folksonomy is a tuple F := (U, T , R, Y) where U, T and R are finite sets, whose elements are respectively called users, tags and resources, and Y is a ternary relation between them, i. e., Y ⊆ U × T × R, whose elements are tag assignments (posts). When a user adds a new or existing resource to a folksonomy, it could be helpful to recommend him/her some relevant tags.</p><p>TRS usually take the users, resources and the ratings of tags into account to suggest a list of tags to the user. According to <ref type="bibr">[15]</ref>, a TRS can briefly be formulated as a system that takes a given user u ∈ U and a resource r ∈ R as input and produces a set T (u, r) ⊂ T of tags as output.</p><p>Jäschke et al. in <ref type="bibr">[10]</ref> define a post of a folksonomy as a user, a resource and all tags that this user has assigned to that resource. This work slightly modifies this definition in the sense that it restricts the set of tags to the tags used simultaneously by a user in order to tag a resource.</p><p>There are some simple but frequently used TRS <ref type="bibr">[10]</ref> based on providing a list of ranked tags extracted from the set of posts connected with the current annotation.</p><p>-MPT (Most Popular Tags): For each tag t i , the posts with t i are counted and the top tags (ranked by occurrence count) are utilized as recommendations. -MPTR (Most Popular Tags by Resource): The number of posts in which a tag occurs together with r i is counted for each tag. The tags occurring most often together with r i are then proposed as recommendations. -MPTU (Most Popular Tags by User): The number of posts in which a tag occurs together with u i is counted for each tag. The tags occurring most often together with u i are then proposed as recommendations. -MPTRU (Most Popular Tags by Resource or User): The number of posts in which a tag occurs together either with r i or u i is counted for each tag. The tags occurring most often together with either r i or u i are taken as recommendations.</p><p>Our hypothesis is that the introduction of a learning system is expected to improve their performance of these systems. These are the key points of the system:</p><p>-The training set depends on each test post and it is specifically built for each of them. Section 4 explains the way of building the initial training set and the example representation. -Several training sets are built according to different kinds of collaboration among users and resources, performing post selection adapting several scoring measures and penalizing oldest posts. Afterwards all of them are compared and evaluated. This approaches are detailed in Section 5. -The learning system adopted was LIBLINEAR <ref type="bibr">[5]</ref>, which provides a probabilistic distribution before the classification. This probability distribution is exerted to rank the tags taking as the most suitable tag the one with highest probability value. The tags of the ranking will be all that appear in the categories of each training set. This entails that some positive tags of a test post might not be ranked. This issue is exposed in depth in Section 6.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Test and Training Data Representation</head><p>This section depicts the whole procedure followed in order to provide a user and a resource with a set of ranked tags. These recommendations are based on a learning process that learns how the users have previously tagged the resources. The core of the method is a supervised learning algorithm based on logistic regression <ref type="bibr">[5]</ref>.</p><p>The traditional approach splits the data into training and test sets at the beginning. Afterwards, a model is inferred using the training set and it is validated thanks to the test set <ref type="bibr">[12]</ref>. In this paper, the methodology used is quite different in the sense that the training and test sets are not fixed. The test set is randomly selected and afterwards an ad hoc training set is provided for each test post. This paper studies different training sets built according to the resource and the user for whom the recommendations are provided.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Definition of the Test Set</head><p>According to the definition of a folksonomy in Section 3, it is composed by a set of posts. Each post is formed by a user, a resource and a set of tags, i.e.,</p><formula xml:id="formula_109">p i = (u i , r i , {t i1 , . . . , t i k })</formula><p>Each post of a folksonomy is candidate to become a test post. Each test post is then turned into as many examples as tags used to label the resource of this post. Therefore, post p i is split into k test examples e 1 = (u i , r i , t i1 ) . . . (3)</p><formula xml:id="formula_110">e k = (u i , r i , t i k )<label>(1</label></formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Definition of the Initial Training Set</head><p>Whichever learning system strongly depends on the training set used to learn. In fact, in order to guarantee a better learning, it would be ideal for the distribution of the categories in both training and test sets to be as similar as possible. Therefore, the selection of an adequate training set is not a trivial task that must be carefully carried out.</p><p>Once the test set is randomly selected, an ad hoc training set is dynamically chosen from the posts posted before the test post.</p><p>The point of departure for building the training set is the set of posts concerning with the resource or the user for which the recommendations are demanded. Once the posts are converted into examples, those examples whose tags have been previously assigned to the resource by the user to whom the recommendations are provided are removed because it has no sense to recommend a user the tags he/she had previously used to label the resource. This section deals with the way of building the initial training set. Next section will explain in depth the way of selecting promising posts through a collaborative approach, using relevance measures for post selection and penalizing oldest posts.</p><p>Let p i = (u i , r i , {t i1 , . . . , t i k }) be a test post. Let R ri be the subset of posts associated to a resource r i and R t ri = {p i /p i ∈ R ri and it was posted bef ore t} Let P ui be the personomy (the subset of posts posted by a user constitutes the so-called personomy) associated to a user u i and P t ui = {p i /p i ∈ P ui and it was posted bef ore t} Therefore, the training set associated to p i is formed by </p><formula xml:id="formula_111">U R di ui,ri = {P d ui ∪ R d ri }\{p j /p j = (u i , r i ,</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">Example Representation</head><p>Now we will explain the way of transforming the post into a computable form understandable for a machine learning system. Therefore, we have to define the features which characterize the examples as well as the class of each example. The features which characterize the examples are the tags previously used to tag the resource in the folksonomy. Hence, each example will be represented by a vector V of size M (the number of tags of the folksonomy) where v j ≥ 1 if and only if t j was used to tag the resource before and 0 otherwise, where j ∈ 1, . . . , M . The class of an example will be the tag the user has tagged the resource with at this moment.</p><p>Let us represent the training set of Example 2.</p><p>Example 3 As an illustration of how to represent a example, let us represent example e 61 of Example 2. The class of e 61 is t 2 , which is its corresponding tag. The features are t 1 and t 3 , since the resource r 1 of e 61 was also tagged before by t 1 in p 1 and p 3 and by t 3 in p 4 . The representation of example e 61 is then {1, 0, 1, 0, 0}.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4">Simple Feature Selection</head><p>An additional proposal to improve the representation is also adopted, since removing redundant or non-useful features which add noise to the system is usually helpful to increase both the effectiveness and efficiency of the classifiers. The example representation based on tags as features makes possible a simple feature selection in the training set. This selection consists of keeping just those tags which represent the test set. Obviously, this is possible just in case the information about the resource of the test post is considered for building the training set, which is the case here. This approach is based on the fact that in a linear system, as the one adopted here, the weights of the features that neither represent the test post nor contribute to obtain the ranking for this post. Therefore, they could be considered as irrelevant features beforehand. This fact can be assumed only for a particular test post. </p><p>In the folksonomy represented in Example 1, resource r 2 does not have any tag assigned before instant d 2 , then its representation is an empty set of features. Analogously, resource r 1 has only been tagged before instant d 3 with t 1 , particularly in instant d 1 by user u 1 , then it is represented only by feature t 1 . The instant d 6 in which the resource r 1 was tagged deserves special attention. Since this resource has been tagged before d 6 with t 1 and t 3 , then both tags are included in its representation. Besides, in example e 61 when the category is t 2 , the tag t 3 is also added because it is a tag assigned in the same instant. In the same way, in example e 62 when the category is t 3 , the tag t 2 is included, since it is a tag assigned in the same instant.</p><p>Reducing such representation to the tags of the test post, the results of this new approach is </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Post Selection</head><p>This section copes with the way of selecting promising posts to be included as examples for the machine learning system. The selection of such posts is carefully carried out taking into account several issues. Let us expose the outline of the process now that will be discussed in depth later. Firstly, only posts that satisfy certain collaborative conditions will be the candidates to add to the initial training set. Secondly, every candidate is scored according to certain measure of relevance. Thirdly, such relevance is penalized depending on the time that the post was posted with regard to the test post. Finally, once a ranking of the candidates is established according to such scoring measure with the correspondent penalization, the most relevance ones will be the posts that will form the final training set. Therefore, the choice of the training set for a given test post is reduced to define the criteria the posts must satisfy to be included in the training set.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1">Collaborative conditions</head><p>Several approaches are proposed to introduce collaboration among users and resources. The effect over the training set is the presence of additional posts carefully selected according to the collaboration among users and/or resources.</p><p>The collaborative conditions can be the following:</p><p>-Collaboration using resources • Take the tags in the posts (contained in the training set described in Section 4.2) that were assigned to the resource r i of the test post p i . Let be this set T ri . • Take the posts (contained in the training set described in Section 4.2) that contain the tags of T ri . • Add such posts to the training set described in Section 4.2 (U R di ui,ri ). Hence, the training set is formed by the posts of U R di ui,ri ∪ T ri . -Collaboration using users</p><p>• Take the tags in the posts (contained in the training set described in Section 4.2) that were assigned by the user u i of the test post p i . Let be this set T ui . • Take the posts (contained in the training set described in Section 4.2) that contain the tags T ui . • Add such posts to the training set described in Section 4.2 (U R di ui,ri ). Hence, the training set is formed by the posts of U R di ui,ri ∪ T ui . -Collaboration using both resources and users by union</p><p>• Take the tags in the posts (contained in the training set described in Section 4.2) that were assigned to the resource r i of the test post p i , that is the set T ri , and that were assigned by the user u i of the test post p i , that is the set T ui . • Take the posts (contained in the training set described in Section 4.2) that contain the tags of T ri ∪ T ui</p><p>• Add such posts to the training set described in Section 4.2 (U R di ui,ri ). Hence, the training set is formed by the posts of U R di ui,ri ∪ T ri ∪ T ui . -Collaboration using both resources and users by intersection</p><p>• Take the tags in the posts (contained in the training set described in Section 4.2) that were assigned to the resource r i of the test post p i , that is the set T ri , and that were assigned by the user u i of the test post p i , that is the set T ui . • Take the posts (contained in the training set described in Section 4.2) that contain the tags of T ri ∩ T ui • Add such posts to the training set described in Section 4.2 (U R di ui,ri ). Hence, the training set is formed by the posts of U R di ui,ri ∪ (T ri ∩ T ui ).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2">Relevance measures</head><p>Once the set of candidates is obtained, they will be scored according to several measures in order to select the most relevant ones. The following scoring measures have been applied before to perform feature selection. Here, they will be adapted to select posts, that, in fact, they are examples instead of features. All of them depend on two parameters, which will be defined before presenting them. The parameter a will be the number of tags that certain post p j shares with the post of test p i (in its representation through the resource of the post as described in Section 4.3) and the parameter b will be the number of tags that certain post p j has (again in its representation through the resource of the post as described in Section 4.3), but the post of test p i does not.</p><p>-From Information Retrieval (IR), document frequency df <ref type="bibr">[21]</ref> and F 1 <ref type="bibr">[22]</ref>.</p><p>-A family of measures coming from Information Theory (IT). These measures consider the distribution of the words over the categories. One of the most widely adopted <ref type="bibr">[18]</ref> is the information gain (IG), which takes into account either the presence of the word in a category or its absence, whereas others are the expected cross entropy for text (CET ) or χ 2 <ref type="bibr">[17]</ref>. They are all defined in terms or probabilities which, in turn, are defined from the parameters mentioned above. -Those which quantify the importance of a feature f in a category c by means of evaluating the quality of the rule f → c, assuming that it has been induced by a Machine Learning (ML) algorithm <ref type="bibr">[18]</ref> (in this paper changing feature by example/post). Some of these measures are based on the percentage of successes and failures of the applications of the rules as, for instance, the Laplace measure (L) which slightly modifies the percentage of success and the difference (D). Other measures that deal with the number of examples of the category in which the feature occurs and the distribution of the examples over the categories are, for example, the impurity level (IL) <ref type="bibr">[20]</ref>. Some other variants of the foresaid measures studying the absence of the feature in the rest of the categories have also been adopted <ref type="bibr">[18]</ref>, leading respectively to the L ir , D ir and IL ir measures.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3">Recent posts and most relevant posts</head><p>A TRS that provides the most on-fashion folksonomy tags would be desirable. This suggest emphasizing the most recent posts more than the oldest ones. For this purpose, a penalizing function is applied to the score granted by a measure. However, some measures reaches negative values, then an increasing function that guaranties a positive value should be applied before the penalizing function in order to keep the ranking the measure gives. The option adopted will be to use the arc tangent and to apply a translation of π 2 . Then, the penalizing functions will be of the form</p><formula xml:id="formula_114">1 (1+ t d ) e ,</formula><p>where t is the time the post was posted, d is the time unit and e is a parameter that controls the penalizing degree. Therefore, if m is the score granted by a measure, the final score granted to each post will be</p><formula xml:id="formula_115">arctan(m) + π 2 • 1 (1 + t d ) e</formula><p>Once the ranking of the posts is established, it is necessary to define a cutoff for selecting the most relevant ones. Some statistics in a folksonomy show that for a given test post its training set might contain either too few posts or too many ones. Both extreme situations are detrimental for the machine learning systems. Applying a percentage of posts to select the most relevant ones avoids neither having too few posts nor to many ones. The alternative used in this paper consists of applying an heuristic able to considerably reduce the posts selected when initially there are too many of them and also able to slightly reduce the posts selected when initially there are too few of them. If n is the original number of posts, such heuristic is defined as follows</p><formula xml:id="formula_116">f loor(2 • n) 0.75</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Learning to Recommend</head><p>The key point of this paper is to provide a ranked set of tags adapted to a user and a resource. Therefore, it could be beneficial to have a learning system able to rank the tags and to indicate the user which tag is the best and which one is the worst for the resource. Taking into account this fact, a preference learning system can not be applied since that kind of methods yield a ranking of the examples (posts) rather than a ranking of categories (tags) <ref type="bibr">[11]</ref>.</p><p>As the input data are multi-category, a system of this kind is expected to be used. However, these systems do not provide a ranking. They can be adapted to produce a partial ranking in the following way: It is possible to take the labels they return and to place them first as a whole and to place the rest of the labels also as a whole afterwards. Obviously, this approach does not establish an order among the labels they recommend but it orders all those labels it returns as a whole with regard to the labels it does not provide.</p><p>The system we need must provide a global ranking of labels. Therefore a multi-label system could be used, but again they need an adaptation to deal with ranking problems. In fact, some multi-label classification systems perform a ranking and then they obtain the multi-label classification <ref type="bibr">[26]</ref>. Hence, it is possible to obtain a ranking directly from them.</p><p>Elisseeff and Weston <ref type="bibr">[6]</ref> propose a multi-label system based on Support Vector Machines (SVM), which generates a ranking of categories. The drawback is that the complexity is cubic and although they perform an optimization to reduce the order to be quadratic, they admit that such complexity is too high to apply to real data sets.</p><p>Platt <ref type="bibr">[19]</ref> uses SVM to obtain a probabilistic output, but just for a binary classification and not for multi-category. A priori one might think about performing as many binary classification problems as the number of tags (categories) that appear in the training set. The problem would turn them into decide if a post is tagged with certain tag or not. But this becomes unfeasible since we are talking of about hundreds of thousands of tags.</p><p>With regard to the problem of tag recommendation, Godbole and Sarawagi in <ref type="bibr">[8]</ref> present an evolution of SVM based on extending the original data set with extra features containing the predictions of each binary classifier and on modifying the margin of SVMs in multi-label classification problems. The main drawback is that they perform a classification rather than a ranking.</p><p>In this framework, LIBLINEAR ([5] and <ref type="bibr">[7]</ref>) is an open source library<ref type="foot" target="#foot_58">3</ref> which is a recent alternative able to accomplish multi-category classification through logistic regression, providing a probabilistic distribution before the classification.</p><p>This paper proposes to use this probability distribution to rank the tags, taking as most suitable tag the one with the highest probability value. In the same sense the most discordant tag will be the one with the lowest probability.</p><p>This work uses the default LIBLINEAR configuration after a slight modification of the output. The evaluation in this case takes place when a resource is presented to the user. Then, a ranking of tags (the tags of the ranking will be all which appear in the categories of the training set) is provided by the learning model.</p><p>If such resource has not been previously tagged, the ranking is generated according to a priori probability distribution. It consists of ranking the tags of the user according to the frequency this user has used them before. Therefore, no learning process is performed in this particular case.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7">Performance Evaluation</head><p>So far, no consensus about an adequate metric to evaluate a recommender has been reached <ref type="bibr">[10]</ref>. Some works do not include quantitative evaluation <ref type="bibr">[28]</ref> or they include it partially <ref type="bibr">[16]</ref>. However, the so called LeavePostOut or LeaveTagsOut proposed in <ref type="bibr">[15]</ref> and <ref type="bibr">[10]</ref> sheds light on this issue. They pick up a random post for each user and they provide a set of tags for this post based on the whole folksonomy except such post. Then, they compute the precision and recall <ref type="bibr">[12]</ref> as follows</p><formula xml:id="formula_117">precision(T ) = 1 |D| (u,r)∈D |T + (u, r) ∩ T (u, r)| |T (u, r)| (7) recall(T ) = 1 |D| (u,r)∈D |T + (u, r) ∩ T (u, r)| |T + (u, r)| (<label>8</label></formula><formula xml:id="formula_118">)</formula><p>where D is the test set, T + (u, r) are the set of tags user u has assigned to resource r (positive tags) and T (u, r) are the set of tags the system has recommended to user u to assign to resource r. The F 1 measure could be computed from them as</p><formula xml:id="formula_119">F 1 = 1 |D| (u,r)∈D 2|T + (u, r) ∩ T (u, r)| |T (u, r)| + |T + (u, r)|<label>(9)</label></formula><p>The evaluation adopted in this paper consists of computing the F 1 , but just considering at most the first five tags of the ranking. Notice that such kind of evaluation quantifies the quality of a classification rather than the quality of a ranking.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="8">Experiments</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="8.1">Data Sets</head><p>The experiments were carried out over the ECML PKDD Dicovery Challenge 2009 datasets <ref type="foot" target="#foot_59">4</ref> . This work studies the T ask 1: Content-Based Tag Recommendations and T ask 2: Graph-Based Recommendations of the 2009 Challenge. The test dataset of the former contains posts, whose user, resource or tags are not contained in the post-core at level 2 of the training data whereas the latter assures that the user, resource, and tags of each post in the test data are all contained in the training data's post-core at level 2.</p><p>The post-core at level 2 is got through cleaning dump and removing all users, tags, and resources which appear in only one post. This process is repeated until convergence and got a core in which each user, tag, and resource occurs in at least two posts.</p><p>The tags were cleaned by removing all characters which are neither numbers nor letters from tags. Afterwards,those tags which were empty after cleaning or matched one of the tags imported, public, systemimported, nn, systemunfiled were removed.</p><p>The cleaned dump contains all public bookmarks and publication posts of BibSonomy<ref type="foot" target="#foot_60">5</ref> until (but not including) 2009-01-01. Posts from the user dblp (a mirror of the DBLP Computer Science Bibliography) as well as all posts from users which have been flagged as spammers have been excluded.</p><p>To make the experiments, the datasets of the Tasks 1 and 2 were split into 2 different datasets. The former is made up by bookmark posts whereas the latter by bibtex posts. These sets will be respectively called bm09 no core and bt09 no core for Task 1 and bm09 core and bt09 core for Task 2.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="8.2">Discussion of results</head><p>This section deals with the experiments carried out. A binary representation of the examples was empirically chosen. Hence, the value of a feature will be 1 if this feature appears in the example and 0 otherwise. For each one, several tag sets are provided depending on the parameters described before:</p><p>-The four ways of collaboration: resource, user, union and intersection.</p><p>-The twelve measures for selecting relevant posts.</p><p>-The penalizing degree of the oldest posts. Several values were checked. Those are 0, 0.0625, 0.25 and 1. The performance corresponding to Tasks 1 and 2 can be seen in the two last rows of Table <ref type="table" target="#tab_1">2</ref>.</p><p>Table <ref type="table" target="#tab_3">3</ref> shows the effect of including collaboration, post selection and a penalization of the oldest posts. All the experiments carried out allow to conclude that the collaboration slightly improves the performance of the recommender. Particularly, the collaboration using resources and using both resources and users by intersection offer the best results with regard to the collaboration using users and using both resources and users by union. Furthermore, collaboration by intersection grants the best results. Including measures to select promising posts improves the recommender. Although the behavior among them is quite similar, measures coming from the Information Theory field together with those based on the impurity level provide the best results. The former seem to be more adequate to the core data whereas the latter improve the results of the no core data. The effect of time differs from one collection to another. The best results are reached without taking into account the time (parameter e = 0) for the bookmark collections, either core or no core versions. However, it seems that penalizing oldest posts improves the performance for the bibtex collections, either core or no core versions.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="9">Conclusions</head><p>This work proposes a TRS based on a novel approach which learns to rank tags from previous posts in a folksonomy using a logistic regression based system. The TRS includes several ways of collaboration among users and resources. It also includes a selection of promising posts using scoring measures and penalizing the oldest ones.</p><p>The collaboration using intersection of tags that both users assign and resources have improves the performance of the recommender with regard to other types of collaboration. Selecting posts using scoring measures makes the recommender provide best tags, although in general the behavior of all of them is quite similar. However, the Information Theory measures offers best results for the core data and the impurity level measures do it for the no core data. Finally, penalizing oldest posts improves the results for the bibtex collections, but it does not obtain satisfactory results for bookmarks collections.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Content-and Graph-based Tag Recommendation: Two Variations</head><p>Johannes Mrosek, Stefan Bussmann, Hendrik Albers, Kai Posdziech, Benedikt Hengefeld, Nils Opperman, Stefan Robert and Gerrit Spira FH Gelsenkirchen, Department of Computer Sciences, Neidenburger Strasse 43, 45877 Gelsenkirchen, Germany, mrosek@internet-sicherheit.de, stefan.bussmann@gmx.net, joschgg@gmx.de, kaip@gmx.net, hengefeld@web.de, nilsoppermann@web.de, FSMARINE@gmx.de, gerrit.spira@gmx.de</p><p>Abstract. We describe two variants of our approach to tackle the task 1 &amp; 2 of the ECML PKDD Discovery Challenge 2009 where each contenter had to identify up to 5 tags for each resource of a given set of either bibtex-like references to publications or bookmarks. The quality of the results was measured against the tags that users of the data source (www.bibsonomy.org) had originally assigned to the resources (F1 measure). In our approach, we either generate tags (from the content of the given resource data or after crawling additional resources) or we request tags from tagging services. We call each of this tag sources a tag recommender. We then combine the results of the tag recommenders based on weighting factors. The weighting factors are determined experimentally by comparing generated and expected tags based on the available training data. This general idea is also used for the graph-based approach required to solve task 2. Here again, the final tag recommendations are computed from the individual results of the different tag-recommending algorithms. In the preliminary result list, we ranked second for task 1 (Group 2) and nineth for task 2 (Group 1).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Key words: content-based graph-based tag recommendation bibsonomy 1 Preliminaries</head><p>Assigning tags to resources can be an effective instrument to organize an information space. Users interacting with this information space may utilize the tags to identify relevant resources or groups of resources. User-driven tag assignment is a popular way to ensure at least some sort of tag quality 1 . Often, the users assigning tags are supported by algorithmic tag recommendation. This may be as simple as offering auto-complete fields for entering tags that display tags already assigned by users, or it may be based on a full-fledged analysis of the resource the user wants to tag. This analysis may present a set of automatically identified tags to the user. Within the later context, compare <ref type="bibr">[1]</ref>, we describe two variants of our approach to content-based tag recommendation which we applied to task 1 and 2 of the ECML PKDD Discovery Challenge 2009. We acted as two teams (with completely independent implementations), each team deploying its own variation of the overall approach -and each with different success. In the following we first describe the solution and results of group 1<ref type="foot" target="#foot_62">2</ref> for tasks 1. This will be followed by a brief presentation of the differences of the solution for task 1 of group 2<ref type="foot" target="#foot_63">3</ref> and their results. We conclude with details of the solution for task 2 of group 1.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Content-based Tag Recommendation</head><p>This attempt on content-based tag recommendation uses different sources to generate tag candidates. These candidates are combined on the basis of appropriate weighting factors which have been assigned to the particular sources ex ante. The first kind of source are web services, which offer information for known resources. This information contains tags which already have been assigned to those resources. In this case we query del.icio.us and citeulike.org. The second kind of source are the resources themselves. In case of bookmarks the content of the according websites is crawled and analyzed. For bibtex entries the information contained in the bibtex table is taken into account. So tag candidates are determined without using any external service. The last source is also a web service called tagthe.net. It already has an engine, which recommends tags for a given URL.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Harvesting Tag Candidates</head><p>The first step to a tag recommendation is the accumulation of candidates and additional information from the several sources. For every URL in the bookmark table, the del.icio.us service is queried. It can be accessed via a feed. The URL md5 hash of the resource is inserted into a URL-pattern. When the resulting URL is accessed, all available information is returned. It includes a count which specifies how often the URL was posted, and a list of top tags assigned to it. Each tag is supplemented with an information about the frequency with witch it has been assigned to the given resource. In some cases it is possible to find tag candidates for bibtex entries at citeulike.org. The description or misc field of the bibtex table often contain a citeulikeid. With this id the according citeulike page can be called and crawled for the assigned tags. These tags are already classified as "tag", "publisher" and "author". Numerical information like a count cannot be obtained. A similar kind of data is provided by the service tagthe.net. It generates classified tags for a submitted URL automatically and independent of the document format behind it. The available categories are "tag", "author", "person" and "location". Like citeulike.org, tagthe.net returns no numerical information. Anyway this service can be seen as a backup system, because it can provide tag candidates for every URL. Thus it is called for all URLs in the bookmark and the bibtex table. Unfortunately there is no information about the algorithmics of the service available.</p><p>In addition to the citeulike-ids the bibtex table contains further interesting information. The title, journal and description fields contain words that can be potentially used as concise tags. Hence after comparison with a stop word list all remaining words in these fields are interpreted as tag candidates.</p><p>The last source for tag candidates are the websites behind the URLs in the bookmark table. After crawling and parsing the site's source code the words are counted and checked against a stop word list. Furthermore they are classified by the location of their appearance like in "meta keywords", "meta description", "title" and "body".</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Selection Tags from the Candidates</head><p>The recommendation system has a hierarchical structure. This means there is one meta recommender which relies on several separate source recommenders, one for each source described in section 1.1. These recommenders provide up to 20 tag candidates for a queried content id. Every tag candidate is complemented by a score the according recommender assigns to it. This score can vary between 0 and 1. How the recommender constitutes that score depends on the source and is described in section 1.3. After the source recommenders have made their suggestions, the meta recommender takes all of the intermediate results and determines the actual tags. Therefor it combines a candidate's frequency of appearance in all sources k with the score s i (1 ≤ i ≤ k) provided by the source recommenders. s i is already influenced by a weighting factor for the individual sources. How this factor is determined will be described in section 1.4. The final score s for a tag candidate can be determined by the formula</p><formula xml:id="formula_120">s = k • (s 1 + s 2 + ... + s k ).<label>(1)</label></formula><p>Note that s can be a value greater than 1. When the meta recommender has calculated these aggregated scores for all candidates, they are ordered by this new information. The five tags with the highest scores are selected as recommended tags. In order to prohibit recommendation of tags with a very low score, an optional filter can be set. This helps to enhance the precision.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3">Calculating the Individual Scores</head><p>As already stated the calculation of the single scores differs depending on the source. The reason for this is the additional information, which is provided besides the tags. Every source offers a different kind of information. To calculate a del.icio.us score s delicious for a tag, the number of posts n which assigned it to the resource is devided by the total number p of posts for the resource at del.icio.us. After that the result is multiplied by the weight w delicious constituted for del.icio.us.</p><formula xml:id="formula_121">s delicious = w delicious • n p<label>(2)</label></formula><p>The scores for tagthe.net and citeulike candidates are determined as follows: For each of the provided categories a weighting factor c categorie is defined: c tag = 1.0, c author = 0.3, c publisher=0.2 , c location = 0.2, c person = 0.1. The score s tagthe/cite of each tag candidate is the result of the product of the source's weight w tagthe/cite and the categorie's weight c categorie . In one case there is an exception from this practice. As can be detected in the training data the tag "juergen" appears pretty often. So this special tag is always handled with a score of 1.0. Otherwise all scores are determined by</p><formula xml:id="formula_122">s tagthe/cite = w tagthe/cite • c categorie .<label>(3)</label></formula><p>The tag candidates generated from the bibtex entries are treated in a similar way. For some fields a weighting factor c f ield is set: c title = 0.55, c journal = 0.25, c description = 0.2. Only tags from the title field are used as a suggestion for the meta recommender. These candidates get a higher score if they also appear in the journal or description field. The individual scores for these fields are added to the initial title-score of 0.55. The scoring process for the crawled content of a website is a little more complex. It is based on the information about the location of appearance as described in section 1.1. Three steps are passed until the final score is determined. In the first step a tag's frequency of appearance n location in the individual locations is counted. This value is weighted by a factor c location which is related to the location's importance. E.g. a word from the title or the keyword-list is rather a good tag than one from the body. All counts are multiplied by their weighting factors get summed up to a kind of weighted frequency n wf req for the whole resource</p><formula xml:id="formula_123">n wf req = n title • c title + n description • c description + n keywords • c keywords + n body • c body . (<label>4</label></formula><formula xml:id="formula_124">)</formula><p>If a tag appears at different locations, its score is raised in step two. This takes into account the fact that such a tag is a good tag in most cases. To raise the score, n wf req is multiplied by a factor f cat which is related to the number of categories the tag appeared in.</p><formula xml:id="formula_125">f cat =        1, 1 categorie 1.5, 2 categories 3, 3 categories 5, 4 categories<label>(5)</label></formula><p>The modified weighted frequency n * wf req is determined by</p><formula xml:id="formula_126">n * wf req = n wf req • f cat .<label>(6)</label></formula><p>In the last step, n * wf req is normalized to the common score interval [0, 1]. Therefor every n wf req is divided by the highest n wf req which is reached for the crawled resource. The result is a score s cc for a tag from the crawled content. The number of tags which are returned to the meta recommender is limited. Only the 20 tags with the highest scores are chosen.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.4">Calculating the Source Weights</head><p>A key point in this two-stage recommendation approach is the calculation of suitable weighting factors for the different sources. Their tag-quality varies in a wide range. Thus handling the candidates of the individual recommenders as if they were of similar quality leads to bad results. As a foundation for the weighting factor calculation the postcore-2 of the training dataset was used. For every source the set of tag assignments it can supply was determined and an according tas file was generated. E.g. to calculate a factor for citeulike, a tas file with all content ids, which are associated with bibtex entries containing a citeulike-id, was generated. Corresponding to the tas files a result file was created for every source. This result was independent of the meta recommender and all the other sources. The result files were measured with f1-score against the tas files created before. The first attempt was to use the f1 score a result file achieved as the weighting factor for the according source recommender. However when testing the complete recommendation process against postcore-2, experiments with various weighting factors pointed out a better choice. Using the precision value which has been computed for the various result files to get the f1-score, provides better overall results. Thus all the weighting factors for the particular sources assumed in the meta recommender match the precision the respective source recommender reaches for the subset of tag assignments it can supply.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.5">Results</head><p>Table <ref type="table" target="#tab_0">1</ref> displays the results each recommender achieved for his own subset of the postcore-2 training data. The intermediate results in the first five rows show recall, precision and f1-score for every source recommender. The weighting factor later used for the meta recommender is the particular precision as described in section 1.4. Row five and six present the results of the meta recommender for the training data. The first is generated without a filter whereas the second one uses a filter to avoid tag recommendations with very low scores as described in section 1.2. The table shows that del.icio.us and citeulike produce good tag candidates. Recommendations made by other sources have a much lower quality. The meta recommenders result for the test dataset is far below the result of the training data set.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.6">Conclusion</head><p>The comparison of the meta recommender's results for training and test data leads to the conclusion that the choice of sources was suboptimal. As the training results were computed on the postcore-2 the web services provided tags with a high quality. Every resource was posted in bibsonomy at least two times. So the probability that it is contained in the web services' databases as well, was pretty good. Thus the decrease in score might be explainable by the unpopularity of the resources in the test dataset. In conclusion the two stage approach is a good attempt for popular resources. To raise the result quality for unpopular resources too, further sources with previously assigned tags have to be added.  Harvesting Tags: In addition to citeulike and del.icio.us, group 2 implemented a tag recommender using Google Scholar for bibtex entries. Using the title data only, links to the resource itself or similar resources were harvested. From the first ten entries three were selected by counting identic words in the titles of the referenced documents (ignoring certain stop words). In the competition, the Google scholar tag recommender harvested roughly 60.000 links for 24.000 bibtex data entries. Subsequently, the Web content crawler of group 2 was used to obtain candidate tags from the three selected resources. Furthermore, the web content crawler was able (to a certain extend) to parse PDF documents in addition to HTML documents. Also the implementation of the citeulike tag recommender differed: a recent database dump of the citeulike data was used which made it possible to search for DOI-ids or URLs directly and to determine the overall frequency of tags in the citeulike data set. TagTheNet was not used. A straighforward tag recommender (called Data set recommender below) that analysed the relevant fields of the data entries complemented the set of recommenders.</p><p>Selecting Tags: No additional filtering for small scores was used. If less than 5 tags were harvested, the remaining slots for tags were filled by randomly drawing tags from the 30 tags most often used in bibsonomy <ref type="foot" target="#foot_64">4</ref> . The weights for a recommender could be different for bibtex and bookmark context. The weights for the recommenders were chosen manually, based on experiences while experimenting with the training data. Tags recommended by more than one recommender were rated significantly higher by multiplying their score with a factor which depends exponentially on the number of recommendations. <ref type="foot" target="#foot_65">5</ref>Results: The results of group 2 for task 1 are presented in table 2. Additional quantitative data are shown in table <ref type="table" target="#tab_3">3</ref>.</p><p>Please note that due to a persistent problem with the CiteULike tag recommender, it is not listed in the training data and haven't been used in the actual competition. Therefore, the relevant F1-score for the ranking in the competition is 0.180 (second place). Some more details on the impact of the different tag recommenders will be presented at the workshop. 3 Graph-Based Tag Recommendation</p><p>The graph-based approach adapts the same hierarchical recommender structure the content-based approach is based on <ref type="foot" target="#foot_66">6</ref> . This means that there is also one meta recommender which combines the results of subordinated recommenders. Instead of using external sources, these recommenders implement different algorithms. Every algorithm processes bibsonomy's graph structure and the contained user information in a different way to generate a set of tag recommendations.</p><p>Like for the content-based approach every tag is valuated with a suitable weighting factor. This weighting factor also depends on the quality of the subordinated recommender. It is determined in the same way as for the source recommenders in Task 1 (cmp. section 1.4). The used benchmark data set contains the whole postcore-2. For processing the final test data set, three different algorithms were used to supply the meta recommender. All of these algorithms have a different focus on evaluating the graph-structure. The specific operation methods are presented in sections 2.1 to 2.3.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Tag by Resource</head><p>The first algorithm selects tags based on the resource information. Therefor all tags which have already been assigned to a queried resource are determined.</p><p>Additionally each tag's frequency of appearance n local in combination with the resource is counted. To calculate the score s tr , this value is taken and divided by the number n post the resource was posted. The result is multiplied by the weighting factor w tr for this recommender. Like in section 1, the recommenders precision for postcore-2 is used for this purpose.</p><formula xml:id="formula_127">s tr = w tr • n local n post<label>(7)</label></formula><p>If two tags reach the same score, the one with the higher frequency of appearance in the entire postcore is preferred.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Tag by User</head><p>The second algorithm recommends and scores tags with a special relation to the user. Algorithmically this approach is very similar to the one described in section 2.1. At first the set of tags is identified which the user who makes a post has assigned to other posts previously. The frequency of usage for the tags n local is also determined. From this set of tags the n local -value for the most used one is taken as a reference value n max . The score s tu this recommender calculates for a tag is as follows</p><formula xml:id="formula_128">s tu = w tu • n local n max<label>(8)</label></formula><p>The weighting factor w tu is determined as usual (cmp. Section 2.1).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Tag by User Similarity</head><p>The last algorithm determines tags, which have been used by similar users. The similiarity between two users is defined over the number of equal resources posted by them.</p><p>Therefor the first task is the determination of similarities between all users. These are calculated by the Tanimoto-Score under consideration of the posted resources. These resources are used as comparison criterion for different users.</p><p>In the recommending process the top five of the similar users are identified. The tags they assigned to the posted resource are accumulated and sorted by their frequency of appearance n count in the posts of the top five users group. As a reference value the maximum frequency of appearance n max of a tag in the same group is utilized. The final score s us for a tag generated by this recommender is calculated with the following formula:</p><formula xml:id="formula_129">s us = w us • n count n max<label>(9)</label></formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4">Results</head><p>The results for Task two are arranged the same way as the results for the contentbased approach in task one. The recall, precision and f1-score values for the training datatset are shown individually for every algorithm. Furthermore results are visualized for meta recommenders which use different combinations of the subordinated recommenders. A filter like the one described in 1.2 is tested too.</p><p>The recommendations for the test dataset have been made by a meta recommender which uses all of the three subordinated recommenders and a filter. They are a good deal worse than the results for the training dataset. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.5">Conclusion</head><p>Like for task 1 there is a dramatic decrease of the f1-score in the test. In this case the explanation can be found in the composition of the training data files. When evaluating the methods for task 2 with the training data, the post, a recommendation is made for, is not excluded from the dataset. Thus the tags of the post are definitely in the training dataset. This increases the probability that the expected tags are recommended. As a consequence, the scores are higher as for the test data which are not included in post-core 2.</p><p>In conclusion, the evaluation of the training data is not valid, but the algorithm works correct for the test data.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Final remarks</head><p>The competition offered an excellent test bed for our approach to tag recommendation. Though we are pretty happy with the results obtained, much work remains to be done:</p><p>-setting the weights for the different recommenders could be improved (iterative algorithmic optimisation; cleaning/choosing training data), -relations to similar resources in the bibsonomy data set (identified for example with the help of the recommended tags or based on Web crawling / Google scholar results) could be explored: if, for a given resource x, a similar resource y in the bibsonomy data sets is identified, tag candidates can be drawn from resource y (either directly or via further graph-based or recursive similarity-based analysis), -fine-tuning of the individual tag recommenders, for example by coupling the Web crawler to the tag database, to name just a few open topics. We want to thank the organizers<ref type="foot" target="#foot_67">7</ref> and to express our hope that this stimulating and exciting competition will be continued in the future and that some of our results and suggestions may help others to develop further improved tag recommenders.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Social tagging is intended to make resources increasingly easy to discover and recover over time. Discovery enables users to find new content of their interest shared by other users. This social indexing gives a promising index quality because it is done by human beings, who understand the content of the resource, as opposed to software, which algorithmically attempts to determine its meaning. Moreover, it is done collectively among users, that is, it uses a collective human intelligence as an index extractor. Recovery enables a user to recall content that was discovered before. It should be easier because the tags are both originated by, and familiar to, its primary users. However, Golder et al. <ref type="bibr">[9]</ref> identify three major problems with the current social tagging systems: polysemy, synonymy, and level variation. The first two inherit the problems of natural language, while the third one refers to the phenomenon of users tagging content at different levels of abstraction. Other problems are dealing with word forms, nouns in singular, nouns in plural, abbreviations, and misspelled words.</p><p>To direct users towards the consistency of the tags, the system usually has a service that assists users in the tagging process, by automatically recommending an appropriate set of tags. The service is a mediated suggestion system, that is, the service does not apply the recommended tags automatically, rather it suggests a set of appropriate tags and allows the user to select tags from the set they find appropriate. Moreover, the tag recommendation can serve many purposes such as consolidating the vocabulary across the users, giving a second opinion what a resource is about and, the important thing, increasing the success of searching because of the consistency <ref type="bibr">[14]</ref>.</p><p>In practice, the standard tag recommenders are services that recommend the most popular tags used for either a particular resource or a whole system. There are other methods proposed from a diversity of approaches to recommend tags from user-created tags (folksonomy) such as information retrieval <ref type="bibr">[23,</ref><ref type="bibr">28]</ref>, graph-based approaches <ref type="bibr">[11]</ref>, collaborative filtering <ref type="bibr">[14]</ref>, machine learning <ref type="bibr">[10,</ref><ref type="bibr">15]</ref>. Recently, people consider textual contents associated to the resources as sources of candidate tags to improve the performance of tag recommenders. For example, Xu et al. <ref type="bibr" target="#b196">[31]</ref> suggest content-based (and context-based) tags based on analysis and classification of the tagged content and context. This not only solves the cold start problem, but also increases the tag quality of those objects that are less popular. Tatu et al. <ref type="bibr">[29]</ref> use natural language processing tools to extract important terms (nouns, adjectives and named entities) from the textual contents. They conclude that the understanding of the contents improves the quality of the tag recommendations.</p><p>In this paper, we also consider the textual contents associated to resources as sources of candidate tags to improve the performance of the tag recommender in the social tagging system. To achieve this goal, we propose a two-level learning hierarchy of concept based keyword extraction as a tag recommendation method. Firstly, the method extracts concepts, which can be considered as a set of related words, using nonnegative matrix factorization (NMF) from training document collections using a two-level learning hierarchy: at the lower level the method extracts concepts and concept-document relationships using usercreated tags. Having these relationships, the method populates the concepts with terms existing in textual contents of resources at the higher level. Next, the tag recommender finds the relevant concepts to a given resource and then scales terms of the resource based on their occurrences in the concepts. The terms having the highest scores are set as keywords and recommended as tags. Incorporating the user-created tags to extract the hidden concept-document relationships distinguishes the two-level from the one-level learning version, which extracts concepts directly from terms existing in textual contents. The main advantage of this approach is that NMF algorithm decomposes more compact document representations. Also, the concept extraction from textual contents is handled by nonnegative least squares algorithm which is much more efficient than NMF algorithm. Therefore, the two-level learning hierarchy approach is not only more efficient but also more reliable because it uses tags created by users who understand the content of documents. Moreover, the approach may have richer vocabularies because it can combine vocabularies from tag space and content space. Our experiment shows that a multi-concept approach, which considers more than one concept for each resource, improves the f-measure values of a single-concept approach, which takes into account just the most relevant concept, about 10%. Moreover, the experiments also prove that the proposed two-level learning hierarchy has f-measure values 13% better than one of the one-level version.</p><p>The rest of the paper is organized as follows: Section 2 discusses a concept extraction method using nonnegative matrix factorization and our proposed two-level learning hierarcy method. Section 3 describes the existing keyword extraction methods and the proposed concept based keyword extraction methods. In Section 4, we describe a tag recommendation algorithm which combines keywords, which are extracted by the keyword extraction methods, with usercreated tags in training data. In Section 5, we show our experiments and results. We conclude and give a summary in Section 6.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Concept Extraction</head><p>Many researchers are trying to address questions about concepts and, in this section, we consider one of them that defines the concepts as a set of related terms. These definitions are proposed and used by some researchers such as <ref type="bibr">[21]</ref> or <ref type="bibr">[27]</ref>. They use clustering methods to extract the concepts from training document collections. Formal concept analysis (FCA) <ref type="bibr">[5,</ref><ref type="bibr">26]</ref> and Latent semantic analysis (LSA) <ref type="bibr">[3,</ref><ref type="bibr">7]</ref> are other methods to perform this task .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">A One-Level Learning Hierarchy for Concept Extraction</head><p>There are some disadvantages of singular value decomposition (SVD) to extract concepts from a document collection as used by LSA. Its negative values make a semantic interpretation difficult. What we would really like to say is that a concept is mostly concerned with some subset of terms, but any semantic interpretation is difficult because of these negative values. To circumvent this problem, a new method which maintains the nonnegative structure of original documents has been proposed. The method uses nonnegative matrix factorization (NMF) <ref type="bibr">[17]</ref> rather than SVD to extract the concepts from document collections.</p><p>Let V be a m × n term-by-document matrix whose columns are document vectors and a positive integer k &lt; min(m, n). In this paper, we use NMF to extract concepts from the term-by-document matrix V . NMF problem is how to find a nonnegative m × k matrix W and a nonnegative k × n matrix H to minimize the functional <ref type="bibr">[2]</ref>:</p><formula xml:id="formula_130">min W,H f (W, H) = 1 2 m i=1 n j=1 V ij − (W H) ij 2 subject to W ia ≥ 0, H aj ≥ 0, ∀i, a, j .<label>(1)</label></formula><p>The constrained optimization problem above is convex on either W or H, but not on both, hence realistic possible solutions usually correspond to local minima.</p><p>The product W H is called a nonnegative matrix factorization of V , although V is not necessarily equal to the product W H. Clearly the product W H is an rank-k approximation to V . An appropriate decision on the value of k is critical in practice, but the choice of k is very often problem dependent. In most cases, however, k is usually chosen such that k &lt;&lt; min(m, n).</p><p>The most popular approach for the NMF problem is the multiplicative update algorithm proposed by Lee and Seung <ref type="bibr">[18]</ref>. To either overcome shortcomings related to convergence properties or to speed up this algorithm, researchers have proposed modifications of the algorithm or even created new ones <ref type="bibr">[2]</ref>. In general, the algorithms can be divided into three general classes: multiplicative update algorithms <ref type="bibr">[18,</ref><ref type="bibr">19]</ref>, gradient descent algorithm <ref type="bibr">[6,</ref><ref type="bibr">12]</ref>, and alternating least squares algorithms <ref type="bibr">[20,</ref><ref type="bibr">24]</ref>.</p><p>Because all elements of the matrix W and H are nonnegative, we can interpret them immediately as following: Each column of W corresponds to a set of related terms called concepts and each element w ia of matrix W represents the degree to which term i belongs to concept a. Each element h aj of matrix H represents the degree to which document j is associated to concept a. Next, we call this type of concept extraction as an one-level learning hierarcy method.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">A Two-Level Learning Hierarchy for Concept Extraction</head><p>In case where the training documents are accompanied by user-created keywords, it is a good idea to incorporate the valuable information in learning process. For this reason, we propose a new learning scheme that uses the keywords for extracting concepts from a document collection. The learning scheme consists of two-level learning hierarchy. At the lower level, concepts and concept-document relationships are discovered using the user-created keywords. Having these relationships, the concepts are populated by terms existing in textual contents of documents at higher level. We expect this method to be successful because the hidden document structures are discovered using keywords collectively created by users. Another advantage of this approach is that NMF algorithm uses more compact document representations. On the other hand, the concept extraction from textual contents is handled by nonnegative least squares algorithm which is much more efficient than NMF algorithm. Therefore, this two-level learning hierarchy approach is not only more efficient but also more reliable because it uses tags created by users who understand the content of documents. Moreover, the approach may have richer vocabularies because it can combine vocabularies from tag space and content space. The detail algorithm of this method is described in Algorithm 1.</p><p>Algorithm 1 A two level learning hierarchy for concept extraction 1: Let V be the tag by document matrix, and X be the term by document matrix 2: Find the tag by concept matrix W and the concept by document matrix H from V = W H using nonnegative matrix factorization (see Section 2.1) to minimize the functional:</p><formula xml:id="formula_131">f (W, H) = 1 2 V − W H 2</formula><p>3: Find the term by concept matrix T from X = T H using nonnegative least squares algorithm, e.g. <ref type="bibr">[2]</ref>:</p><p>-Solve for T in matrix equation HH T T T = HX T -Set all negative elements in T to 0</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Concept Based Keyword Extraction</head><p>Keyword extraction is the task of automatically selecting a small set of important, topical terms within the textual content of a document. The fact that the keywords are extracted means that the selected terms are present in the document <ref type="bibr">[16]</ref>. In general, the task of automatically extracting keywords can be divided into two stages:</p><p>1. Selecting candidate terms in the document 2. Filtering out the most significant ones to serve as keywords and rejecting those that are inappropriate There are various methods proposed for selecting candidate terms. The first one is n-gram extraction, that is, extracting uni-, bi-, or tri-grams, removing those that begin or end with a stop word <ref type="bibr">[8]</ref>. Another one is more linguistically oriented using natural language processing (NLP) method such as NP-chunker or part-of-speech (PoS) <ref type="bibr">[13]</ref>. Filtering uses either simple statistics, where a weighting schema is applied to rank words accoding to their score <ref type="bibr">[1,</ref><ref type="bibr">25]</ref>, or machine learning, where the ranking function is defined by a statistical model derived from training set with manually assigned keywords <ref type="bibr">[13,</ref><ref type="bibr">22,</ref><ref type="bibr" target="#b195">30]</ref>.</p><p>In this section, we propose a machine learning based filtering method, that is, a method that uses concepts extracted from textual contents of documents. The method finds the relevant concepts to a given document and then scales terms of the document based on their occurrences in the concepts. The terms having the highest scores are set as keywords. The method can be considered as unsupervised learning when we use the one-level learning hierarchy. It means that the method does not need labeled data for the training process. Moreover, the method becomes a supervised method when the user-created tags are used in learning process for the two-level learning hierarchy approach. Two variants of the concept based keyword extraction method are described in detail in the following sections.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Single-Concept Based Keyword Extraction</head><p>The two-level learning hierarchy extracts concepts from a training document collection. Having these concepts, the single-concept based keyword extraction method finds the most relevant concept to a given document and then scales the candidate terms existing in the document based on their occurrence in the concept. The relevance of a concept c with a document d is calculated using the following cosine distance measure:</p><formula xml:id="formula_132">rel(d, c) = d T V c d V c<label>(2)</label></formula><p>The detail algorithm of this approach is described in Algorithm 2. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Multi-Concept Based Keyword Extraction</head><p>The multi-concept based keyword extraction assumes that a document may contain more than one relevant concept. The detail algorithm of the multi-concept based keyword extraction method is described in Algorithm 3. If this is the case then the tag space based recommenders are suggested. The collaborative recommender is used if a given user has profiles in system. Otherwise, the most popular tag by resource method is used as tag recommender.</p><p>If the resource appears for the first time then the recommender examines the content of the resource using the concept based keyword extraction algorithm. Boosting the extracted tags if they have been used by the user before. If neither any tags nor any keywords are suggested then the most popular tags in the training data are recommended. Using the user-created tags aims to direct the standardization and consistency of supplied tags, while using the tags extracted from textual contents intend especially to overcome the cold start problem.</p><p>Algorithm 4 A hybrid tag recommendation algorithm </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Experiment</head><p>We apply our proposed recommender methods (Algorithm 4) for ECML PKDD Discovery Challenge 2009<ref type="foot" target="#foot_69">1</ref> . The task of the competition requires the development of a content-based tag recommendation method for BibSonomy<ref type="foot" target="#foot_70">2</ref> , a web based social bookmarking system that enables users to tag both web pages (bookmark) and scientific publications (bibtex). The organizers of the competition made available a training set of examples consisting of the resources accompanied with their user-created tags. A testing data will be provided in order to evaluate proposed recommenders. Each bookmark is described by its URL, a description of the URL that usually is the title of the web page and an extended description of the bookmark supplied by the user. Each bibtex is associated with values of bibtex fields such as title, author, booktitle, journal, series, volume, number, etc. BibtexKey, bibtexAbstract, URL, and description of the publication can be specified. Some statistics of the data are shown in Table <ref type="table" target="#tab_0">1</ref>. In our experiment, we use textual contents associated to each resource as content of the resources. For the bookmark, the contents are the description of the URL and the extended description. Title and abstract are textual contents associated to the bibtex. A bookmark is identified by its URL address (url hash) attribute and a bibtex by its title (simhash1 ) attribute. Therefore, a document, bookmark or bibtex, is represented by the description given to the document by all users that bookmarked the document.</p><p>Let D be a testing data set, consisting of |D| examples (r i , T i ), i = 1...|D|. Let T i be the set of tags created by users for a resource r i and P i be the set of tags predicted by a recommender for a resource r i . The precision, recall, and F-measure for recommender f on testing data set D is calculated as follows:</p><formula xml:id="formula_133">Precision = 1 |D| |D| i=1 |T i ∩ P i | |P i | Recall = 1 |D| |D| i=1 |T i ∩ P i | |T i | F − Measure = 1 |D| |D| i=1 2 |T i ∩ P i | |P i | + |T i |</formula><p>We perform our experiment in java platform and use Lucene<ref type="foot" target="#foot_71">3</ref> for creating the tag-by-resource matrix and the term-by-resource matrix. The other processes are conducted on the Weka <ref type="foot" target="#foot_72">4</ref> framework, an open source machine learning software.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1">Experiment Settings</head><p>For each of the method of our experiment the settings we used to run them are described as following:</p><p>Concept-based keyword extraction . For creating the term-by-resource matrix, resources are parsed and a dictionary of terms is created using a standard word tokenization method. The terms are words, special characters are removed, and Snowball Porter stemming and standard stop words of English and German are applied. Finally, the term-by-resource matrix is created using a term frequency weighting scheme.</p><p>Extracting concepts from the term-by-resource matrix is an important step to find keywords from new resources. The optimal number of concepts (k), which captures most concepts in the training document collection, remains difficult to find. The method that is usually used for a practical purpose is a heuristic approach. However, because of the memory usage, simulations are usually conducted on the maximum number of concepts that can be extracted. In our experiment, we extract 200 concepts for the training document collection. For this task, we use the nonnegative double SVD initialization method <ref type="bibr">[4]</ref> that conducts no randomization and the projected gradient method <ref type="bibr">[20]</ref> that converges to a local minimum. We expect these combining methods leads to converge to a unique solution with a minimum error.</p><p>There is another parameter α that should be optimized in the two-level learning hierarchy approach. The parameter reflects the portion of the tag space and the content space as sources of tags for the recommender. In our experiment, we set the parameter α = 0.25 for the single-concept method and α = 0.05 for the multi-concept method, which are the optimal values we get using a heuristic method.</p><p>Collaborative recommendation <ref type="bibr">[14]</ref> . For a given tag-by-user matrix X, a given user u, a given resource r, and integer k and n, the set T (u, r) of n recommended tags is calculated by:</p><formula xml:id="formula_134">T (u, r) = argmax n t∈T v∈N k u sim(X u , X v )δ(v, t, r)<label>(3)</label></formula><p>where</p><formula xml:id="formula_135">N k u is k nearest neighbors of u in X, δ(v, t, r) = 1 if (v, t, r</formula><p>) ∈ f olksonomy and 0 else. Therefore, the only parameter to be tuned is the number of neighbors k. For that, multiple runs where performed where k incremented until a point where no more improvement in the results were observed.</p><p>Most popular tags by resource . For a given resource we count how many posts a tag occur together with that resource. We use tags that occur most often together with that resource as recommendation.</p><p>Most popular tags . For each tags we count in how many posts it occurs. We then use tags that occur most often as recommendation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2">Single-vs. Multi-Concept Method</head><p>Fig. <ref type="figure" target="#fig_0">1</ref> shows the performances of the single-and multi-concept based keyword extraction on testing data. From Fig. <ref type="figure" target="#fig_0">1</ref>, we can calculate that recall, precision, and f-measure of the multi-concept approach are, on average, 10%, 15% and 12%. The recall is likely to increase when the number of recommended tags gets bigger, while the precision is reduced for the bigger numbers of tags. Fig. <ref type="figure" target="#fig_0">1</ref> also shows the performance of the single-concept approach in the similar pattern and its f-measure is, on average, 11%. From both curves, we conclude that the multi-concept approach, which assumes that a resource may contain more than one concept, improves f-measure of the single-concept method, on average, 10%. The improvement occurs in all numbers of recommended tags. These results verify that associating of resources with more than one concept gives better performance than just considering the main concept of resources. In other words, some minor concepts of a resource should also be examined for getting the better performance of the keyword extraction.  In this section, we examine the performance of our proposed two-level learning hierarchy approach compared to the one-level version. Fig. <ref type="figure" target="#fig_4">2</ref> shows the performance of the one-level learning hierarchy multi-concept based keyword extraction and the two-level learning hierarchy multi-concept based keyword extraction. From the figure, we see that the two-level learning method has better recall, precision and f-measure. Its f-measure values, on average, are 13% better than one of the one-level learning approach. The detailed recall, precision, and f-measure values of the optimal performance of the two-level learning hierarchy are given in Table <ref type="table" target="#tab_1">2</ref>.   </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Summary</head><p>In this paper, we propose a two-level learning hierarchy concept based keyword extraction method for task1 of ECML PKDD Discovery Challenge 2009, that is, a content-based tag recommendation. The tag recommendation method explores tags from textual contents of resources using concepts existing in the textual contents of the resources. A multi-concept approach, which considers more than one concept for each resource, improves the performance of a singleconcept approach, which only considers the most relevant concept. Moreover, our experiment demonstrates that the proposed two-level learning hierarchy method outperforms the common one-level learning approach for all performance measures, e.g. recall, precision, and f-measure.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>STaR: a Social Tag Recommender System 1 Introduction</head><p>The coming of Web 2.0 has changed the role of Internet users and the shape of services offered by the World Wide Web. Since web sites tend to be more interactive and user-centric than in the past, users are shifting from passive consumers of information to active producers. By using Web 2.0 applications, users are able to easily publish content such as photos, videos, political opinions, reviews, so they are identified as Web prosumers: producers + consumers of knowledge. One of the forms of user-generated content (UGC) that has drawn more attention from the research community is tagging, which is the act of annotating resources of interests with free keywords, called tags, in order to help users in organizing, browsing and searching resources through the building of a sociallyconstructed classification schema, called folksonomy <ref type="bibr">[18]</ref>. In contrast to systems where information about resources is only provided by a small set of experts, collaborative tagging systems take into account the way individuals conceive the information contained in a resource <ref type="bibr">[19]</ref>. Well-known example of platforms that embed tagging activity are Flickr<ref type="foot" target="#foot_73">1</ref> to share photos, YouTube<ref type="foot" target="#foot_74">2</ref> to share videos, Del.icio.us<ref type="foot" target="#foot_75">3</ref> to share bookmarks, Last.fm<ref type="foot" target="#foot_76">4</ref> to share music listening habits and Bibsonomy<ref type="foot" target="#foot_77">5</ref> to share bookmarks and lists of literature. Although these systems provide heterogeneous contents, they have a common core: once a user is logged in, she can post a new resource and choose some significant keywords to identify it. Besides, users can label resources previously posted from other users. This phenomenon represents a very important opportunity to categorize the resources on the web, otherwise hardly feasible. The act of tagging resources from different users is the social aspect of this activity; in this way tags create a connection among users and items. Users that label the same resource by using the same tags could have similar tastes and items labeled with the same tags could have common characteristics. Many would argue that the power of tagging lies in the ability for people to freely determine the appropriate tags for a resource without having to rely on a predefined lexicon or hierarchy <ref type="bibr">[11]</ref>. Indeed, folksonomies are fully free and reflect the user mind, but they suffer of the same problems of unchecked vocabulary. Golder et. al. <ref type="bibr">[5]</ref> identified three major problems with current tagging systems: polysemy, synonymy, and level variation. Polysemy refers to situations where tags can have multiple meanings: for example a resource tagged with the term turkey could indicate a news taken from an online newspaper about politics or a recipe for Thanksgiving' Day. When multiple tags share a single meaning we refer to it as synonymy. In collaborative tagging systems we can have simple morphological variations (for example we can find 'blog', 'blogs', 'web log', to identify a common blog) but also semantic similarity (like resources tagged with 'arts' versus 'cultural heritage'). The third problem, called level variations, refers to the phenomenon of tagging at different level of abstraction. Some people can annotate a web page containing a recipe for roast turkey with the tag 'roastturkey' but also with a simple 'recipe'.</p><p>In order to avoid these problems, in the last years many tools have been developed to facilitate the user in the task of tagging and to aid the tag convergence <ref type="bibr">[4]</ref>: these systems are know as tag recommenders. When a user posts a resource in a Web 2.0 platform, a tag recommender suggests some significant keywords to label the item following some criteria to filter out the noise from the complete tag space. This paper presents STaR (Social Tag Recommender system), a tag recommender system developed for the ECML-PKDD 2009 Discovery Challenge. The idea behind our work is that folksonomies create connections among users and items, so we tried to point out two concepts:</p><p>-Resources with similar content could be annotated with similar tags; -A tag recommender needs to take into account the previous tagging activity of users, by weighting more tags already used to annotate similar resources.</p><p>In this work we identify two main aspects in the tag recommendation task: firstly, each user has a typical manner to label resources (for example using personal tags such as 'beautiful', 'ugly', 'pleasant', etc. which are not connected to the content of the item, or simply tagging using general tags like 'politics', 'sport', etc.); next, similar resources usually share common tags: when a user posts a resource r on the platform, our system takes into account how she (if she is already stored in the system) and the entire community previously tagged resources similar to r in order to suggest relevant tags. Next, we develop this model and we tested it on a dataset extracted from BibSonomy.</p><p>The paper is organized as follows. Section 2 analyzes related work. The general problem of tag recommendation is introduced in Section 3. Section 4 explains the architecture of the system and how the recommendation approach is implemented. The experimental section carried out is described in Section 5.1, while conclusions and future works are drawn in last section.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Related Work</head><p>Previous work in the tag recommendation area can be broadly divided into three classes: content-based, collaborative and graph-based approaches.</p><p>In the content-based approach, a system exploits some textual source with Information Retrieval-related techniques <ref type="bibr">[1]</ref> in order to extract relevant unigrams or bigrams from the text. Brooks et. al <ref type="bibr">[3]</ref>, for example, develop a tag recommender system that automatically suggests tags for a blog post extracting the top three terms exploiting TF/IDF scoring <ref type="bibr">[14]</ref>. The system presented by Lee and Chun <ref type="bibr">[8]</ref> recommends tags retrieved from the content of a blog using artificial neural networks. The network is trained based on statistical information about word frequencies and lexical information about word semantics extracted from WordNet. The collaborative approach for tag recommendation, instead, presents some analogies with collaborative filtering methods <ref type="bibr">[2]</ref>. In the model proposed by Mishne and implemented in AutoTag <ref type="bibr">[12]</ref>, the system suggests tags based on the other tags associated with similar posts in a given collection. The recommendation process is performed in three steps: first, the tool finds similar posts and extracts their tags. All the tags are then merged, building a general folksonomy that is filtered and reranked. The top-ranked tags are suggested to the user, who selects the most appropriate ones to attach to the post. TagAssist <ref type="bibr">[16]</ref> improves the AutoTags' approach performing a lossless compression over existing tag data. It finds similar blog posts and suggests a subset of the associated tag through a Tag Suggestion Engine (TSE) which leverages previously tagged posts providing appropriate suggestions for new content. In <ref type="bibr">[10]</ref> the tag recommendations task is performed through a user-based collaborative filtering approach. The method seems to produce good results when applied on the user-tag matrix, so they show that users with a similar tag vocabulary tend to tag alike. The problem of tag recommendation through graph-based approaches has been firstly addressed by Jäschke et al. in <ref type="bibr">[7]</ref>. They compared some recommendation techniques including collaborative filtering, PageRank and FolkRank. The key idea behind FolkRank algorithm is that a resource which is tagged by important tags from important users becomes important itself. The same concept holds for tags and users, thus the approach uses a graph whose vertices mutually reinforce themselves by spreading their weights. The evaluation showed that FolkRank outperforms other approaches. Schmitz et al. <ref type="bibr">[15]</ref> proposed association rule mining as a technique that might be useful in the tag recommendation process. In literature we can find also some hybrid methods integrating two or more approaches (mainly, content and collaborative ones) in order to reduce their typical drawbacks and point out their qualities. Heymann et. al <ref type="bibr">[6]</ref> present a tag recommender that exploits at the same time social knowledge and textual sources. They suggest tags based on page text, anchor text, surrounding hosts, adding tags used by others users to label the URL. The effectiveness of this approach is also confirmed by the use of a large dataset crawled from del.icio.us for the experimental evaluation. A hybrid approach is also proposed by Lipczak in <ref type="bibr">[9]</ref>. Firstly, the system extracts tags from the title of the resource. Afterwards, based on an analysis of co-occurrences, the set of candidate tags is expanded adding also tags that usually co-occur with terms in the title. Finally, tags are filtered and reranked exploiting the information stored in a so-called "personomy", the set of the tags previously used by the user.</p><p>Finally, in <ref type="bibr">[17]</ref> the authors proposed a model based on both textual content and tags associated with the resource. They introduce the concept of conflated tags to indicate a set of related tag (like blog, blogs, ecc.) used to annotate a resource. Modeling in this way the existing tag space they are able to suggest various tags for a given bookmark exploiting both user and document models. They win the previous edition of the Tag Recommendation Challenge.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Description of the Task</head><p>STaR has been designed to participate at the ECML-PKDD 2009 Discovery Challenge <ref type="foot" target="#foot_78">6</ref> . In this section we will firstly introduce a formal model for recommendation in folksonomies, then we will analyze the specific requirements of the task proposed for the Challenge.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Recommendation in Folksonomies</head><p>A collaborative tagging system is a platform composed of users, resources and tags that allows users to freely assign tags to resources. Following the definition introduced in <ref type="bibr">[7]</ref>, a folksonomy can be described as a triple (U, R, T ) where:</p><p>-U is a set of users; -R is a set of resources; -T is a set of tags.</p><p>We can also define a tag assignment function tas: U × R → T . The tag recommendation task for a given user u ∈ U and a resource r ∈ R can be finally described as the generation of a set of tags tas(u, r) ⊆ T according to some relevance model. In our approach these tags are generated from a ranked set of candidate tags from which the top n elements are suggested to the user.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Description of the ECML-PKDD 2009 Discovery Challenge</head><p>The 2009 edition of the Discovery Challenge consists of three recommendation tasks in the area of social bookmarking. We compete for the first task, contentbased tag recommendation, whose goal is to exploit content-based recommendation approaches in order to provide a relevant set of tags to the user when she submits a new item (Bookmark or BibTeX entry) into Bibsonomy.</p><p>The organizers make available a training set with some examples of tag assignment: the dataset contains 263,004 bookmark posts and 158,924 BibTeX entries submitted by 3,617 different users. For each of the 235,328 different URLs and the 143,050 different BibTeX entries were also provided some textual metadata (such as the title of the resource, the description, the abstract and so on).</p><p>Each candidate recommender is evaluated by comparing the real tags (namely, the tags a user adopts to annotate an unseen resource) with the suggested ones. The accuracy is finally computed using classical IR metrics, such as Precision, Recall and F1-Measure (Section 5.1).</p><p>By analyzing the aforementioned requirements, we designed STaR thinking at a prediction task rather than a recommendation one. Consequently, we will try to emphasize the previous tagging activity of the user, also looking for connections and patterns among resources. All these decisions will be thoroughly analyzed in the next section describing the architecture of STaR.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">STaR: a Social Tag Recommender System</head><p>STaR (Social Tag Recommender) is a content-based tag recommender system, developed at the University of Bari. The inceptive idea behind STaR is to improve the model implemented in systems like TagAssist <ref type="bibr">[16]</ref> or AutoTag <ref type="bibr">[12]</ref>. Although we agree with the idea that resources with similar content could be annotated with similar tags, in our opinion Mishne's approach presents two important drawbacks: 1. The tag reranking formula simply performs a sum of the occurrences of each tag among all the folksonomies, without considering the similarity with the resource to be tagged. In this way tags often used to annotate resources with a low similarity level could be ranked first. 2. The proposed model does not take into account the previous tagging activity performed by users. If two users bookmarked the same resource, they will receive the same suggestions since the folksonomies built from similar resources are the same.</p><p>We will try to overcome these drawbacks, by proposing an approach based on the analysis of similar resources capable also of weighting more the tags already selected by the user during her previous tagging activity. Figure <ref type="figure" target="#fig_0">1</ref> shows the general architecture of STaR. The recommendation process is performed in four steps, each of which is handled by a separate component.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Indexing of Resources</head><p>Given a collection of resources (corpus), a preprocessing step is performed by the Indexer module, which exploits Apache Lucene<ref type="foot" target="#foot_79">7</ref> to perform the indexing step. As regards bookmarks we indexed the title of the web page and the extended description provided by users. For the BibteX entries we indexed the title of the publication and the abstract. Let U be the set of users and N the cardinality of this set, the indexing procedure is repeated N + 1 times: we build an index for each user (Personal Index ) storing the information on her previously tagged resources and an index for the whole community (Social Index ) storing the information about all the resources previously tagged by the community.</p><p>Following the definitions presented in Section 3.1, given a user u ∈ U we define P ersonalIndex(u) as:</p><formula xml:id="formula_136">P ersonalIndex(u) = {r ∈ R|∃t ∈ T : tas(u, r) = t}<label>(1)</label></formula><p>where tas is the tag assignment function tas: U × R → T which assigns tags to a resource annotated by a given user. SocialIndex represents the union of all the user personal indexes:</p><formula xml:id="formula_137">SocialIndex = N i=1 P ersonalIndex(u i ) (2)</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Retrieving of Similar Resources</head><p>At the end of the preprocessing step STaR is able to take into account users requests. Every user interacts with STaR by providing information about a resource to be tagged. In the Query Processing step the system acquires data about the user (her language, the tags she uses more, the number of tags she usually uses to annotate resources, etc.) before processing (through the elimination of not useful characters and punctuation) and submitting the query against the SocialIndex stored in Lucene. If the user is recognized by the system since it has previously tagged some other resources, the same query is submitted against her own PersonalIndex, as well. We used as query the title of the web page (for bookmarks) or the title of the publication (for BibTeX entries). In order to improve the performances of the Lucene Querying Engine we replaced the original Lucene Scoring function with an Okapi BM25 implementation <ref type="foot" target="#foot_80">8</ref> . BM25 is nowadays considered as one of the state-of-the art retrieval models by the IR community <ref type="bibr">[13]</ref>.</p><p>Let D be a corpus of documents, d ∈ D, BM25 returns the top-k resources with the highest similarity value given a resource r (tokenized as a set of terms t 1 . . . t m ), and is defined as follows:</p><formula xml:id="formula_138">sim(r, d) = m i=1 n r ti k 1 ((1 − b) + b * lengthr avgLengthr ) + n r ti * idf (t i )<label>(3)</label></formula><p>where n r ti represents the occurrences of the term t i in the document d, length r is the length of the resource r and avgLength r is the average length of resources in the corpus. Finally, k 1 and b are two parameters typically set to 2.0 and 0.75 respectively, and idf (t i ) represents the inverse document frequency of the term t i defined as follows: where N is the number of resources in the collection and df (t i ) is the number of resources in which the term t i occurs.</p><formula xml:id="formula_139">idf (t i ) = log N + df (t i ) + 0.5 df (t i ) + 0.5<label>(4)</label></formula><p>Given user u ∈ U and a resource r, Lucene returns the resources whose similarity with r is greater or equal than a threshold β. To perform this task Lucene uses both the PersonalIndex of the user u and the SocialIndex. More formally: P ersonalRes(u, q) = {r ∈ P ersonalIndex(u)|sim(q, r) ≥ β} <ref type="bibr">(5)</ref> SocialRes(q) = {r ∈ SocialIndex|sim(q, r) ≥ β} (6) Figure <ref type="figure" target="#fig_4">2</ref> depicts an example of the retrieving step. In this case the target resource is represented by Gazzetta.it, one of the most famous Italian sport newspaper. Lucene queries the SocialIndex and returns as the most similar resources an online newspaper (Corrieredellosport.it) and the official web site of an Italian Football Club (Inter.it). The PersonalIndex, instead, returns another online newspaper (Tuttosport.com). The similarity score returned by Lucene has been normalized.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">Extraction of Candidate Tags</head><p>In the next step the Tag Extractor gets the most similar resources returned by the Apache Lucene engine and produces the set of candidate tags to be suggested, by computing for each tag a score obtained by weighting the similarity score returned by Lucene with the normalized occurrence of the tag. If the Tag Extractor also gets the list of the most similar resources from the user Person-alIndex, it will produce two partial folksonomies that are merged, assigning a weight to each folksonomy in order to boost users' previously used tags.</p><p>Formally, for each query q (namely, the resource to be tagged), we can define a set of tags to recommend by building two sets: candT ags p and candT ags s . These sets are defined as follows:</p><p>candT ags p (u, q) = {t ∈ T |t = T AS(u, r) ∧ r ∈ P ersonalRes(u, q)} (7)</p><formula xml:id="formula_140">candT ags s (q) = {t ∈ T |t = T AS(u, r) ∧ r ∈ SocialRes(q) ∧ u ∈ U }<label>(8)</label></formula><p>In the same way we can compute the relevance of each tag with respect to the query q as: rel p (t, u, q) = r∈P ersonalRes(u,q) n t r * sim(r, q) n t (9) rel s (t, q) = r∈SocialRes(q) n t r * sim(r, q) n t <ref type="bibr">(10)</ref> where n t r is the number of occurrences of the tag t in the annotation for resource r and n t is the sum of the occurrences of tag t among all similar resources.</p><p>Finally, the set of Candidate Tags can be defined as:</p><p>candT ags(u, q) = candT ags p (u, q) ∪ candT ags s (q) <ref type="bibr">(11)</ref> where for each tag t the global relevance can be defined as:</p><formula xml:id="formula_141">rel(t, q) = α * rel p (t, q) + (1 − α) * rel s (t, q)<label>(12)</label></formula><p>where α (PersonalTagWeight) and (1 − α) (SocialTagWeight) are the weights of the personal and social tags respectively. Figure <ref type="figure" target="#fig_10">3</ref> depicts the procedure performed by the Tag Extractor : in this case we have a set of 4 Social Tags (Newspaper, Online, Football and Inter) and 3 Personal Tags (Sport, Newspaper and Tuttosport). These sets are then merged, building the set of Candidate Tags. This set contains 6 tags since the tag newspaper appears both in social and personal tags. The system associates a score to each tag that indicates its effectiveness for the target resource. Besides, the scores for the Candidate Tags are weighted again according to SocialTagWeight (α) and PersonalTagWeight (1 − α) values (in the example, 0.3 and 0.7 respectively), in order to boost the tags already used by the user in the final tag rank. Indeed, we can point out that the social tag 'football' gets the same score of the personal tag 'tuttosport', although its original weight was twice. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4">Tag Recommendation</head><p>The Tag Extractor produces the set of the Candidate Tags, a ranked set of tags with their relevance scores. This set is exploited by the Filter, a component which performs the last step of the recommendation task, that is removing those tags not matching specific conditions: we fix a threshold for the relevance score between 0.20 to 0.25 and we return at most 5 tags. These parameters are strictly dependent from the training data.</p><p>Formally, given a user u ∈ U , a query q and a threshold value γ, the goal of the filtering component is to build recommendation(u, q) defined as follows:</p><p>recommendation(u, q) = {t ∈ candT ags(u, q)|rel(t, q) &gt; γ} <ref type="bibr">(13)</ref> In the example in Figure <ref type="figure" target="#fig_10">3</ref>, setting a threshold γ = 0.20, the system would suggest the tags sport and newspaper.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Experimental Evaluations</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1">Experimental Session</head><p>In this experiment we measure the performance of STaR in the Task 1 of the ECML-PKDD 2009 Discovery Challenge. This experimental evaluation was carried out according to the instructions provided from the organizers of the Challenge 2009. The test set was released 48 hours before the end of the competition. Every participant uploaded a file containing the tag predictions, and for each post only five tags were considered. F1-Measure was used to evaluate the accuracy of recommendations, thus for each post Precision and Recall were computed by comparing the recommended tags with the true tags assigned by the users. The case of tags was ignored and all characters which are neither numbers nor letters were removed. Results are presented in Table <ref type="table" target="#tab_0">1</ref>. STaR finished the ECML-PKDD Discovery Challenge 2009 with an overall F-measure of 13.55. As showed in the table above, exploiting only the first recommended tag the system reaches almost 20% in precision. The value of the recall increases with the number of recommended tags reaching the 13.5% in the fourth and fifth tag. In the future we will perform a more in-depth study in order to compare the predictive accuracy of STaR with different configurations of parameters.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Conclusions and Future Work</head><p>In this paper we presented STaR, a tag recommender designed and implemented to participate to the ECML-PKDD 2009 Discovery Challenge. The idea behind our work was to discover similarity among resources in order to exploit communities and user tagging behavior. In this way our recommender system was able to suggest tags for users and items still not stored in the training set. The experimental sessions showed that users tend to reuse their own tags to annotate similar resources, so this kind of recommendation model could benefit from the use of the user personal tags before extracting the social tags of the community (we called this approach user-based).</p><p>In the future we will implement a methodology to suggest tags when the set of similar items returned by Lucene is empty. The system should be able to extract significant keywords from the textual content associated to a resource (title, description, etc.) that has not similar items, maybe exploiting structured data or domain ontologies. Another issue to investigate is the application of our methodology in different domains such as multimedia environment. In this field discovering similarity among items just on the ground of textual content could be not sufficient. Finally, textual content suffers from syntactic problems like polysemy (a keyword with two or more meanings) and synonymy (two or more keywords with the same meaning). These problems hurt the performance of the recommender. We will try to establish if a semantic document indexing could improve the performance of the recommender.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Collaborative tagging systems or folksonomies have steadily gained popularity in the recent years. Users are free to choose the tags they want to use, and while this may be a main reason behind the popularity of these systems, it is also one of the biggest problems these systems face. As users come up with new tags they forget the tags they used to use, making it difficult to find the previously tagged content. Tag recommendation can help both in search and in keeping the users' tagging practices consistent. Tag recommendation can be defined as the problem of finding suitable tags or labels to a given resource for a given user.</p><p>Tag recommendation can be an important element in a folksonomy as it can help users employ the tags consistently as well as help users to use same tags for similar resources. This can improve searching within the users' own resources as well as the folksonomy.</p><p>We present a method for tag recommendation that combines several baseline methods and collaborative filtering. Combining the results makes use of the past performance of the recommenders.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Tag Recommendation</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Collaborative Filtering for Folksonomies</head><p>Collaborative filtering (CF), a popular method used in recommender systems can be adapted for tag recommendation. The description here is based on <ref type="bibr">[1]</ref>.</p><p>Folksonomy can be understood as a tuple F = (U, R, T, Y ), where U is the set of users, T is the set of tags and R is the set of resources (bookmarks and BibTeX entries in the case of BibSonomy <ref type="bibr">[2]</ref>) and Y ⊆ U × R × T is the tag assignment relation. those users, resources and tags that appear at least in two posts. The test set for this task was known to have the users, resources and tags from this set.</p><p>We processed bookmarks and BibTeX entries identically. The only information extracted from the "bookmark" and "bibtex" tables were the hash values which identified the resources. We used the url hash and simhash1 columns and did not attempt to combine duplicate resources. The url hash considers two resources different if there are any differences in the url, such as a trailing slash.</p><p>To retain a slightly better neighbourhoods for the collaborative filtering approach we used full training set to calculate the neighbourhoods, but removed the tags that could not appear in the results. The difference between this and the post-core at level 2 was that this left several partial posts to the training data.</p><p>No effort was made to separate functional tags (such as "myown" and "toread") from descriptive tags, which are considerably more interesting in tag recommendation.</p><p>Some of the most used tags in BibSonomy are used by a small minority, such as "juergen" (3101 posts, 2 users). In total, in the subset of tags that are contained in the post-core 2 there are 273 tags that have been used at least 100 times by at most 5 people. A measure for the popularity of the tag, which takes into account the number of users of a tag can be defined as</p><formula xml:id="formula_142">popularity(t) = log(N t ) * log(N * t ),<label>(4)</label></formula><p>where N t is the number of times the tag t has been used and N * t is number of users for the tag t.</p><p>This measure can be used to improve tag recommendation methods which would not otherwise give weights to different tags. As can be seen from Table <ref type="table" target="#tab_0">1</ref>, sorting the tags by their "popularity" removes the unlikely tag "zzztosort" while preserving a sensible selection of popular tags.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Results</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Combining Recommendations</head><p>The baseline methods can yield good results on certain users, but they are generally worse than the alternatives. However, combining the baseline results with results from collaborative filtering or other methods can be used to improve the general results. The problem of combining results is in evaluating the trustworthiness of the recommender results.</p><p>In tag recommendation, there are multiple "items" that are recommended, and besides the similarity between the user and the neighbours of the user there are few evident factors that could be used to weight the tags when combining different methods. In our method, we used the training data to predict the recent posts of the users (1-100 posts, but at most 20% of the user's all posts)</p><p>In our approach, we took the arbitrary set of methods shown in Table <ref type="table" target="#tab_1">2</ref> and assigned weights to different tags by calculating the weighted sum over all recommenders using the per-user per-post weighted sum</p><formula xml:id="formula_143">w t := p [t ∈ T ] * 0.9 k f p (5)</formula><p>where f p is the F-measure of the method p ∈ 1, .., 7 on the validation set, and k is the position of the tag in the recommendation. This reduces the weight of the tag slightly so that the methods with smaller F-measure have a better possibility of getting a likely tag in the final results. The final recommendation are the five t ∈ T with the highest w t . Prior to the competition, we performed a test with the training data. The posts were divided into three sets based on the post date. The first 80% was selected to work as a training set, the following 10% as the validation set and the last 10% were used for testing. The method weights were computed from the validation set. The resulting weights were tested on the test set, showing a modest 5% improvement in the F-measure over the best baseline method in the test.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Experiment on the Competition Set</head><p>The weights for the methods were assigned to the users in the competition set by generating recommendations for recent posts with all the methods listed in the previous section. The amount of posts was chosen was up to 100 posts, but at most 20% of the user's all posts. After this, the F-measure for each method was used to generate a mixing profile for each user. Then the recommendations were made for the competition set and these were combined using the equation 5. The results are summarized in Table <ref type="table" target="#tab_3">3</ref>. One of the baselines (resource tags) outperforms the combined result slightly on the competition set. Some of the recommendations, such as "resource tags", can contain very unlikely tags when the resource itself is tagged only a few times and contains unpopular tags; this was not taken into account when combining the recommendations. A possible solution for this problem is to not recommend unlikely (unpopular) tags if the user hasn't used them in the past.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Conclusion</head><p>In these experiments, the weights of the recommenders are based on their past performance, but it is likely that there are several features that can be used to estimate these weights from statistical features of the user, such as the average "popularity" of the user's tags and the number of distinct tags. We would like to study these numbers for correlations. Recommendations by other methods, such as FolkRank <ref type="bibr">[1]</ref> could be added to improve the performance on the dense parts of the data.</p><p>The obtained results were less than stellar; in retrospect, more attention should have been paid to the combining of the results and especially the fact that the results of the recommendations were far from independent. Some method for filtering the results should have been applied, perhaps by modifying the weights for the individual tags by using the information whether the target user has used a certain tag before and how popular the tag is. Simple methods should not be completely neglected, as they can provide useful results for users who do not conform to the tagging practices of the mainline users of the folksonomy.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Discussion</head><p>F-measure works as a performance measure for tag recommendation to a certain extent, but the utility of tag recommendation methods for usability and search within a folksonomy should be confirmed with user tests. Combining different tag recommendation results with different weights at different times may cause the recommendation to feel inconsistent.</p><p>Searching within a folksonomy is sometimes unnecessarily difficult. A part of the problem is that users tend to use only a few tags per post. One improvement for these tagging systems would be to ask for applicability of a set of tags that are similar to the ones user has already chosen. It might make sense to distinguish between the problems of tag prediction, that is, predicting the tags user will choose, and tag recommendation, the problem of finding descriptive tags for a resource.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Factor Models for Tag Recommendation in BibSonomy</head><p>Steffen Rendle and Lars Schmidt-Thieme</p><p>Machine Learning Lab, University of Hildesheim, Germany {srendle,schmidt-thieme}@ismll.uni-hildesheim.de</p><p>Abstract. This paper describes our approach to the ECML/PKDD Discovery Challenge 2009. Our approach is a pure statistical model taking no content information into account. It tries to find latent interactions between users, items and tags by factorizing the observed tagging data. The factorization model is learned by the Bayesian Personal Ranking method (BPR) which is inspired by a Bayesian analysis of personalized ranking with missing data. To prevent overfitting, we ensemble the models over several iterations and hyperparameters. Finally, we enhance the top-n lists by estimating how many tags to recommend.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>In this paper, we describe our approach to task 2 of the ECML/PKDD Discovery Challenge 2009. The setting of the challenge is personalized tag recommendation <ref type="bibr">[1]</ref>. An example is a social bookmark site where a user wants to tag one of his bookmark and the tag recommender suggest the user a personalized list of tags he might want to use for this item. Our approach to this problem is a pure statistical model using no content information. It relies on a factor model related to <ref type="bibr">[2]</ref> where the model parameters are optimized for the maximum likelihood estimator for personalized pairwise ranking <ref type="bibr">[3]</ref>. Furthermore, we use a smoothing method for reducing the variance in the factor models. Finally, we provide a method for estimating how many tags should be recommended for a given post. This method is model independent and can be applied to any tag recommender.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Terminology and Formalization</head><p>We follow the terminology of <ref type="bibr">[2]</ref>: U is the set of all users, I the set of all items/ resources and T the set of all tags. The tagging information of the past is represented as the ternary relation S ⊆ U × I × T . A tagging triple (u, i, t) ∈ S means that user u has tagged an item i with the tag t. The posts P S denotes the set of all distinct user/ item combinations in S:</p><formula xml:id="formula_144">P S := {(u, i)|∃t ∈ T : (u, i, t) ∈ S}</formula><p>Our models calculate an estimator Ŷ for S. Given such a predictor Ŷ the list Top of the N highest scoring items for a given user u and an item i can be calculated by:</p><formula xml:id="formula_145">Top(u, i, N ) := N argmax t∈T ŷu,i,t<label>(1)</label></formula><p>where the superscript N denotes the number of tags to return. Besides ŷu,i,t we also use the notation of a rank ru,i,t which is the position of t in a post (u, i) after sorting all tags by ŷu,i,t : ru,i,t := |{t : ŷu,i,t &gt; ŷu,i,t }|</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Factor Model</head><p>Our factorization model (FM) captures the interactions between users and tags as well as between items and tags. The model equation is given by:</p><formula xml:id="formula_146">ŷu,i,t = f ûu,f • tU t,f + f îi,f • tI t,f<label>(2)</label></formula><p>Where Û , Î, T U and T I are feature matrices capturing the latent interactions.</p><p>They have the following types:</p><formula xml:id="formula_147">Û ∈ R |U |×k , Î ∈ R |I|×k , T U ∈ R |T |×k , T I ∈ R |T |×k</formula><p>Note that this model differs from the factorization model in <ref type="bibr">[2]</ref> where the model equation is the Tucker Decomposition.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Optimization Criterion</head><p>Our optimization criterion is an adaption of the BPR criterion (Bayesian Personalized Ranking) <ref type="bibr">[3]</ref>. The criterion presented in <ref type="bibr">[3]</ref> is derived for the task of item recommendation. Adapted to tag recommendation, the optimization function for our factor model is:</p><formula xml:id="formula_148">BPR-Opt := (u,i)∈P S t + ∈T + u,i t − ∈T − u,i ln σ(ŷ u,i,t + − ŷu,i,t − ) − λ(|| Û || 2 + || Î|| 2 + || T U || 2 + || T I || 2 ) (3)</formula><p>That means BPR-Opt tries to optimize the pairwise classification accuracy within observed posts. Note that it differs from <ref type="bibr">[2]</ref> by optimizing for pairwise classification (log-sigmoid) instead of AUC (sigmoid).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Learning</head><p>The model is learned by the LearnBPR algorithm <ref type="bibr">[3]</ref> which is a stochastic gradient descent algorithm where cases are sampled by bootstrapping. In the following, we show how we apply this generic algorithm to the task of optimzing our model paramaters for the task of tag recommendation. The gradients of our model equation <ref type="bibr">(2)</ref> with respect to the model parameters Θ = { Û , Î, T U , T I } are:</p><formula xml:id="formula_149">∂BPR-Opt ∂Θ = (u,i)∈P S t + ∈T + u,i t − ∈T − u,i ∂ ∂Θ ln σ(ŷ u,i,t + − ŷu,i,t − ) − λ ∂ ∂Θ ||Θ|| 2 ∝ (u,i)∈P S t + ∈T + u,i t − ∈T − u,i −e −(ŷ u,i,t + −ŷ u,i,t − ) 1 + e −(ŷ u,i,t + −ŷ u,i,t − ) • ∂ ∂Θ (ŷ u,i,t + − ŷu,i,t − ) − λΘ</formula><p>That means, we only have to compute the derivations of our model equation ŷu,i,t with respect to each model parameter from Θ = { Û , Î, T U , T I }:</p><formula xml:id="formula_150">∂ ∂ ûu,f ŷu,i,t = tU t,f ∂ ∂ îu,f ŷu,i,t = tI t,f ∂ ∂ tU t,f ŷu,i,t = ûu,f ∂ ∂ tI t,f ŷu,i,t = îi,f</formula><p>These derivations are used in the stochastic gradient descent algorithm shown in figure <ref type="figure" target="#fig_0">1</ref>.</p><p>The method presented so far has the following hyperparameters:</p><formula xml:id="formula_151">-α ∈ R + learning rate -λ ∈ R + 0 regularization parameter -µ ∈ R mean value for initialization of model parameters -σ 2 ∈ R + 0 standard deviation for initialization of model parameters -k ∈ N + feature dimensionality of factorization</formula><p>Reasonable values for all parameters can be searched on a holdout set. The learning rate and the initialization parameters are only important for the learning algorithm but are not part of the optimization criterion or model equation. Usually, the values found for α, µ, σ 2 on the holdout generalize well.</p><p>In contrast to this, the regularization and dimensionality are more important for the prediction quality. In general, when the regularization is chosen properly, the higher the dimensionality the better. In our submitted result, we use an ensemble of models with different regularization and dimensionality. </p><formula xml:id="formula_152">,f ← ûu,f + α " e −d 1+e −d • ( tU t + ,f − tU t − ,f ) + λ • ûu,f " 8: îi,f ← îi,f + α " e −d 1+e −d • ( tI t + ,f − tI t − ,f ) + λ • îi,f<label>"</label></formula><formula xml:id="formula_153">9: tU t + ,f ← tU t + ,f + α " e −d 1+e −d • ûu,f + λ • tU t + ,f " 10: tU t − ,f ← tU t − ,f + α " e −d 1+e −d • −û u,f + λ • tU t − ,f<label>"</label></formula><formula xml:id="formula_154">11: tI t + ,f ← tI t + ,f + α " e −d 1+e −d • îi,f + λ • tI t + ,f " 12: tI t − ,f ← tI t − ,f + α " e −d 1+e −d • − îi,f + λ • tI t − ,f<label>"</label></formula><p>13: end for 14:</p><p>until convergence 15:</p><p>return Û , Î, T U , T I 16: end procedure Fig. <ref type="figure" target="#fig_0">1</ref>. Optimizing our factor model for equation <ref type="bibr">(3)</ref> with bootstrapping based stochastic gradient descent. With learning rate α and regularization λ.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Ensembling Factor Models</head><p>Ensembling factor models with different regularization and dimensionality is supposed to remove variance from the ranking estimates. There are basically two simple approaches of ensembling predictions ŷl u,i,t of l models:</p><p>1. Ensemble of the value estimates ŷl u,i,t :</p><formula xml:id="formula_155">ŷev u,i,t := l w l • ŷl u,i,t<label>(4)</label></formula><p>2. Ensemble of the rank estimates rl u,i,t :</p><formula xml:id="formula_156">ŷer u,i,t := l w l • (|T | − rl u,i,t )<label>(5)</label></formula><p>That means tags with a high rank (low r) will get a high score ŷ.</p><p>Where w l is the weighting parameter for each model. Whereas ensembling value estimates is effective for models with predictions on the same scale, rank estimates are favorable in cases where the ŷ values of the different models have no direct relationship.</p><p>Ensembling Different Factor Models For our factor models the scales of ŷ depend both on the dimensionality and the regularization parameter. Thus we use the rank estimates for ensembling factor models with different dimensionality and regularization. In our approach we use a dimensionality of k ∈ {64, 128, 256} and regularization of λ ∈ {10 −4 , 5 • 10 −5 }. As the prediction quality of all of our factor models are comparable, we have chosen identical weights w l = 1.</p><p>Ensembling Iterations Within each factor model we use a second ensembling strategy to remove variance. Besides the hyperparameters, another problem is the stopping criterion of the learning algorithm (see figure <ref type="figure" target="#fig_0">1</ref>). We stop after a predefined number of iterations (2000) -we have chosen an iteration size of 10 • |S| single draws. In our experiments the models usually converged already after about 500 iterations but in the following iterations the ranking alternates still a little bit. To remove the variance, we create many value estimates from different iterations and ensemble them. I.e. after the first 500 iterations we create each 50 iterations a value estimate for each tag in all test posts and ensemble these estimates with (4). Again there is no reason to favor an iteration over another, so we use identical weights w l = 1. This gives the final estimates for each model. The models with different dimensionality and regularization are ensembled as described above.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Baseline Models</head><p>Besides our factorization model we also consider several baseline models and ensembles of these models. The models we pick as baselines are most-popular by item (mpi), most-popular by user (mpu), item-based knn (knni) and user-based knn (knnu).</p><p>The most-popular models are defined as follows:</p><formula xml:id="formula_157">ŷmpi u,i,t = |{u ∈ U : (u , i, t)}| ŷmpu u,i,t = |{i ∈ I : (u, i , t)}|</formula><p>The k-nearest-neighbour models (knn) are defined as follows:</p><formula xml:id="formula_158">ŷknni u,i,t = (u,i ,t)∈S sim i,i ŷknnu u,i,t = (u ,i,t)∈S sim u,u</formula><p>To measure sim i,i and sim u,u respectively, we first fold/ project the observed data tensor in a two dimensional matrix F U and F I :</p><formula xml:id="formula_159">f I i,t = |{u : (u, i, t) ∈ S}| f U u,t = |{i : (u, i, t) ∈ S}|</formula><p>Based on these simple estimators, a combined post size can be produced by a linear combination:</p><formula xml:id="formula_160"># E u,i := β 0 + β G # G u,i + β U # U u,i + β I # I u,i</formula><p>In our approach we use # E u,i and optimize β on the holdout set for maximal F1. We found that choosing an adaptive length of the recommender list significantly improved the results over a fixed number.</p><p>6 Experimental Results</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.1">Sampling of Holdout Set</head><p>As the test of the challenge was released two days before the submission deadline, we tried to generate representative holdout-sets. We created two test sets, one following the leave-one-post-per-user-out protocol <ref type="bibr">[1]</ref> and a second one by uniformly sampling posts with the constraint that the dataset should remain a 2-core after moving a post into the test set. These two sets were used as holdout sets for algorithm evaluation and hyperparameter selection. In the following, we report results for the second holdout set, because its characteristics (in terms of number of users, items and posts) are closer to the real test set.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.2">Results</head><p>The results of the method presented so far can be found in table 2 and 3. As you can see, the single baseline models result in low quality but ensembles can achieve a good quality. In contrast to this, our proposed factor models generate better recommendations. The best possible ensemble (optimized on test!) of the baselines achieves a score of 0.330 on the challenge set whereas our factor ensemble (not optimized on test) results in 0.345. mpu mpi mp-ens knni knnu knn-ens knn+mp-ens holdout 0.249 0.351 -/0.423 0.401 0.371 -/0.445 -/0.473 challenge 0.098 0.288 0.290/0.317 0.209 0.295 0.293/0.320 0.299/0.330 Fig. <ref type="figure" target="#fig_4">2</ref>. F-Measure quality for the baselines methods. For the ensembles, we report two results: one for an ensemble with identical weights and one with optimal weights that have been optimized on the test! set. For sure this is an optimistic value that might not be found using the holdout split.</p><p>An interesting finding is that the results on the challenge test set largely differs from both of our holdout sets. But as all methods suffer, we assume that the tagging behavior in the challenge test set is indeed different from the one in the training set. Especially, the baseline most-popular-by-user dropped largely from 24.9% to 9.8% -this might indicate that personalization is difficult to achieve on single FM FM-ens FM-ens adaptive list length holdout 0.495 ± 0.002 0.498 0.522 challenge -0.345 0.356 Fig. <ref type="figure" target="#fig_10">3</ref>. F-Measure quality for the factorization methods. Single FM reports the average quality of each factorization model. FM-ens is the unweighted ensemble and finally we report the ensemble with the adaptive list length, i.e. predicting sometimes less than 5 tags.</p><p>the challenge test set using the provided training set. Non-personalized methods or content-based methods could benefit from the difference in both sets. Also methods that can handle temporal changes in the tagging behaviour might improve the scores.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7">Conclusion</head><p>In this paper, we have presented a factor model for the task of tag recommendation. The model tries to describe the individual tagging behavior by four low-dimensional matrices. The model parameters are optimized for the personalized ranking criterion BPR-Opt <ref type="bibr">[3]</ref>. The length of the recommended lists is adapted both to the user and item. Our evaluation indicates that our approach outperforms ensembles of baseline models which are known to give high quality recommendations <ref type="bibr">[1]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Social tagging, aka, folksonomy, is a popular way to organize resources like documents, bookmarks and photos. Resource, tag and user are three essential parts in a social tagging system, a user uses tags to describe resources. Tag suggestion system eases the process of social tagging. It can suggest tags to new resources based on previous tagged resources.</p><p>To promote related research, ECML/PKDD organizes a open contest of tag suggestion systems, named Discovery Challenge 2009 (DC09 in short). In this contest a snapshot of users, documents and tags in the online bookmarking system BibSonomy is provided. Each team trains their suggestion system on the snapshot, and test the performance on the same test dataset. There are 3 tasks in the contest. Task 1 focuses on suggesting tags by the content of the resources, i.e, content-based tag suggestion. Task 2 focuses on suggesting tags by the tripartite links between resources, tags and users, i.e., graph-based tag suggestion. Task 3 puts the suggestion system into real-life situation by integrating it with BibSonomy website, and see which system predicts the user's intention best.</p><p>In this paper, we describe our methods for the three tasks. For Task 1 and 3, we propose a fast tag suggestion method called Feature-Driven Tagging (FDT). FDT indexes tags by features, where feature can be word, resource ID, user ID or others. For each feature, FDT keeps a list of weighted tags, the higher the weight, the more likely the tag is suggested by the feature. For a new resource, each feature in it suggests a list of weighted tags, the suggestions are combined according to the importance of features to get the final suggestion. Compared to other methods, FDT provides suggestions faster, and the speed is only related with the number of features in the resource(number of words in the content).</p><p>For Task 2, we apply two existing methods, most popular tags and FolkRank, for graph-based suggestion. Furthermore, we propose to use a new graph-based ranking model, DiffusionRank, for tag suggestion. The method of "most popular tags" is the simplest collaborative-filtering based methods. It recommends the most popular tags of the resources used by other users. FolkRank is based on PageRank <ref type="bibr">[1]</ref> on user-resource-tag tripartite graph, which was first proposed as a tag suggestion method in <ref type="bibr">[2]</ref>. DiffusionRank was originally proposed for combating web spam <ref type="bibr">[3]</ref>, which has also been successfully used in social network analysis <ref type="bibr">[4]</ref> and search query suggestion <ref type="bibr">[5]</ref>. DiffusionRank is motivated by the heat diffusion process, which can be used for ranking because the activities flow on the graph can be imagined as heat flow, the edge from a vertex to another can be treated as the pipe of an air-conditioner for heat flow. Compared to PageRank, DiffusionRank provides more flexible mechanism to make the ranking scores related to initial values of the vertices, which is important for graph-based tag suggestion.</p><p>The paper is organized as follows. Section 2 formulates the problem of tag suggestion. Section 3 introduces our method for content-based tag suggestion. Section 4 introduces our method for graph-based tag suggestion. Section 5 describes the dataset, experiment settings and the result. Section 6 introduces related work on tag suggestion. Section 7 concludes the paper.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Problem Formulation</head><p>We adopt the model of social tagging proposed by Jaschke et al <ref type="bibr">[2]</ref>. A social tagging data set is defined as a tuple F := (U, T, R, Y ), where U is the set of users, T is the set of tags and R is the set of resources. Y is a ternary relation between U, T and R, Y ⊆ U × T × R. (u, r, t) ∈ Y is called a tag assignment, which means user u assigned the tag t to resource r. A resource r ∈ R can be described with a piece of text, like titles of a paper or user-edited description of a website. We denote the words in the text as {w i }.</p><p>Resources, users and tags form a graph G = (V, E), where V = U R T , and E = {{u, t}, {u, r}, {r, t}|(u, t, r) ∈ Y }. The goal of tag suggestion is to predict the set of tags {t} for a given pair of user and resource (u, r).</p><p>In related literature, social tags are also called folksonomy, the pair of a resource and a user is also called a post.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Content-based Tag Suggestion</head><p>In this section, we propose a content-based tag suggestion method named Feature-Driven Tagging(FDT). Briefly speaking, FDT is a voting model, where each feature in the resource votes for their favorite tags, and the final scores of tags are averaged by the importance of the features. Figure <ref type="figure" target="#fig_0">1</ref> illustrates the tagging procedure of FDT, it consists of 3 steps: feature extraction, feature weighting and tag voting. For a resource with content, FDT first extracts features from the content. Features include but are not limited to words, resource ID and user ID. Then, FDT weights each feature by their importance in the resource, we explain different ways to compute the importance of features later in this section. In the voting step, each feature contributes a weighted list of tags, the higher the weight, the more likely we should suggest the tag. Weight of a tag from different features are combined by the importance of each feature, thus creates the final weighted list of tags. In the tagging process, all parameters are indexed by feature, we do not need to iterate over all tags (as in text categorization approaches) or resources (as in neighborhood-based approaches), so it is called Feature-Driven Tagging.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Feature Extraction</head><p>We extract features from different sources. Word features are extracted from textual content of resources, we use them to capture the relationship between words and tags. For bibtex, the textual content is title + bibtexAbstract + journal + booktitle + annote + note + description; For bookmark, it is description + extended. We also include simhash1 and the user ID of a resource as a feature. The same publication or website share the same simhash1, we use it to capture the tags assigned by other users. We use user ID as a feature so as to model a user's preferences of tagging.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Compute the Importance of Features</head><p>We use two methods to assess the importance of features in a resource. The first and most intuitive one is T F × IDF . T F × IDF is widely used in information retrieval, text categorization and keyword extraction <ref type="bibr">[6]</ref>. We use log-version of T F × IDF , which computes as</p><formula xml:id="formula_161">T F × IDF (f ) = log( n(f ) N + 1) * log( |R| df (f ) + 1)<label>(1)</label></formula><p>where n(f ) is the number of occurrences of f in this resource, and N is the total number of features occurred in this resource. |R| is the total number of resources, df (f ) is the number of resources f has occurred in. The +1 is to avoid zero or negative weights.</p><p>The other method we used is T F × IT F , ITF stands for Inverse Tag Frequency, it computes as follows,</p><formula xml:id="formula_162">T F × IT F (f ) = log( n(f ) N + 1) * log( |T | ntag(f ) + 1)<label>(2)</label></formula><p>where |T | is the total number of tags, and ntag(f ) is the number of tags f has co-occurred with. ITF implies that the more tags a feature co-occurs with, the less specific and important the feature is.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Feature-Tag Correlation</head><p>In FDT, each feature is associated with a weighted list of tags. We denote this as a matrix Θ, where θ i,j is the weight of tag t j to feature f i , the size of Θ is |F | × |T |, F is the set of all features. Although Θ is large, it is extremely sparse, so each feature only associates with a small number of tags.</p><p>We use three different methods to compute Θ offline, they are co-occurrence count(CC), Mutual Information (MI) and χ 2 statistics (χ 2 ). Co-occurrence count is computed by</p><formula xml:id="formula_163">CC(f, t) = n(f, t)/n(t)<label>(3)</label></formula><p>where n(f, t) is the number of co-occurrences of feature f and tag t, and n(t) is the total number of occurrences of tag t. CC is a naive way to find the most important tags for a feature.</p><p>In MI, we model each feature or tag as a binary-valued probabilistic variable, the value of which means occur in a document <ref type="bibr">(1)</ref> or not(0). Then, we can compute the Mutual Information between a feature and a tag by</p><formula xml:id="formula_164">M I(f, t) = f ′ ∈f, f t ′ ∈t, t p(f ′ , t ′ )log( p(f ′ , t ′ ) p(f ′ )p(t ′ ) )<label>(4)</label></formula><p>where f ′ = f means feature f occurs in the resource, and f ′ = f means it doesn't occur, the same is for t ′ . MI computes the shared information between f and t, the higher it is, the more correlated f and t are.</p><p>χ 2 has been used for feature selection in text categorization <ref type="bibr">[7]</ref>, it also find the correlation between a feature and a category, here we use the tag as category. χ 2 is computed as follows,</p><formula xml:id="formula_165">χ 2 (f, t) = N (AD − BC) 2 (A + C)(B + D)(A + B)(C + D)<label>(5)</label></formula><p>where</p><formula xml:id="formula_166">A = n(f, t), B = n(f, t), C = n( f , t, D = n( f , t).</formula><p>After we get Θ by one of the above methods, we make Θ sparse by picking the largest K values in Θ and set other values to 0. We test K = 30000, 50000 and 100000, as K increases, the F1 measure increases. When K &gt; 50000, the F1-measure doesn't change a lot, so we use K = 50000 in all experiments. For each row in Θ, we first find the largest value θ i,max , then set all values in this row to θ i,j = θ i,j /θ i,max . We compare the performance of these 3 methods in the experiment section.</p><p>FDT has low computation complexity when tagging. For a resource with n features, the complexity of tagging is O(nm), where m is the average tags for each feature in Θ. m is usually a small number, in our model it is 4.63 for bibtex and 5.81 for bookmark. Note that the complexity of FDT is not related to the total number of training documents, tags or users. Nearest neighbor methods have to search in the entire training data set, so the complexity is at least O(|R|). Multi-label classifier methods have to train a classifier for each one of tags, so the complexity is at least O(|T |). Furthermore, the model of FDT is related with K, which is around 10 5 , it is small enough to load in the main memory.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Graph-based Tag Suggestion</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Method Preliminaries</head><p>The basic idea of graph-based tag suggestion is to construct a graph with users, resources and tags as vertices and build edges according to user tagging behaviors. After building the graph, we can adopt some graph-based ranking algorithms to rank tags for a specific user and resource. Then the top-ranked tags are recommended to users.</p><p>To describe the graph-based methods more clearly, we first give some mathematical notations. For the folksonomy F := (U, T, R, Y ), we firstly convert it into an undirected tripartite graph G F = (V, E). In G F , the vertices consists of users, resources and tags, i.e., V = U R T . For each tagging behavior of user u assigning tag t to resource r, we will add edges between u , r and t, i.e., E = {{u, r}, {u, t}, {r, t}|(u, t, r) ∈ Y }.</p><p>In G F , we have the set of vertices</p><formula xml:id="formula_167">V = {v 1 , v 2 , • • • , v N } and the set of edges E = {(v i , v j ) |</formula><p>There is an edge between v i and v j }. For a given vertex v i , let N (v i ) be the set of vertices that are neighbors of v i . We have w(v i , v j ) as the weight of the edge (v i , v j ). For an undirected graph, w(v i , v j ) = w(v j , v i ). Let w(v i ) be the degree of v i , and we have</p><formula xml:id="formula_168">w(v i ) = vj ∈N (vi) w(v j , v i ) = vj ∈N (vi) w(v i , v j )<label>(6)</label></formula><p>With the matrix, we can rewrite the Equation <ref type="formula" target="#formula_48">9</ref>as:</p><formula xml:id="formula_169">s = λAs + (1 − λ)p (<label>10</label></formula><formula xml:id="formula_170">)</formula><p>where s is the vector of PageRank scores of vertices, and p is the vector of preferences of vertices.</p><p>A straightforward idea of graph-based tag suggestion is to set preference to the user and resource to be suggested for, and then compute ranking values using PageRank in Eq. <ref type="bibr">(10)</ref>. However, as pointed out in <ref type="bibr">[8]</ref>, this will make it is difficult for other vertices than those with high edge degrees to become highly ranked, no matter what the preference values are.</p><p>Based on above analysis, we described FolkRank as follows. To generate tags for user u and resource r, we have to:</p><p>1. Let s (0) be the stable results of Eq. ( <ref type="formula" target="#formula_49">10</ref>) with p = 1, i.e., the vector composed by 1's. 2. Let s (1) be the stable results of Eq. ( <ref type="formula" target="#formula_49">10</ref>) with p = 0, but p(u) = 1 and p(r) = 1. 3. Compute s := s (1) − s (0) . Therefore, we can rank tags according to their final values in s, where the topranked tags are suggested to user u for resource r.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4">DiffusionRank</head><p>DiffusionRank was originally proposed for combating web spam <ref type="bibr">[3]</ref>, which has also been successfully used in social network analysis <ref type="bibr">[4]</ref> and search query suggestion <ref type="bibr">[5]</ref>. DiffusionRank is motivated by the heat diffusion process, which can be used for ranking because the activities flow on the graph can be imagined as heat flow, the edge from a vertex to another can be treated as the pipe of an air-conditioner for heat flow.</p><p>For a graph G = {V, E}, denote f i (t) is the heat on vertex v i at time t, we construct DiffusionRank as follows. Suppose at time t, each vertex v i receives an amount of heat, M (v i , v j , t, ∆t), from its neighbor v j during a period ∆t. The received heat is proportional to the time period ∆t and the heat difference between v i and v j , namely f j (t) − f i (t). Based on this, we denote M (v i , v j , t, ∆t) as</p><formula xml:id="formula_171">M (v i , v j , t, ∆t) = γ(f j (t) − f i (t))∆t</formula><p>where γ is heat diffusion factor, i.e. the thermal conductivity. Therefore, the heat difference at node v i between time t + ∆t and time t is equal to the sum of the heat that it receives from all its neighbors. This is formulated as:</p><formula xml:id="formula_172">f i (t + ∆t) − f i (t) = vj ∈N (vi) γ(f j (t) − f i (t))∆t<label>(11)</label></formula><p>The process can also be expressed in a matrix form:</p><formula xml:id="formula_173">f (t + ∆t) − f (t) ∆t = γHf (t) (<label>12</label></formula><formula xml:id="formula_174">)</formula><p>where f is a vector of heat at vertices at time t, and H is</p><formula xml:id="formula_175">H(i, j) =    −1 if i = j 0 if (v i , v j ) / ∈ E w(vi,vj ) w(vj ) if (v i , v j ) ∈ E<label>(13)</label></formula><p>If the limit ∆t → 0, the process will become into</p><formula xml:id="formula_176">d dt f (t) = γHf (t)<label>(14)</label></formula><p>Solving this differential equation, we have f (t) = e γtH f (0). Here we could extend the e γtH as</p><formula xml:id="formula_177">e γtH = I + γtH + γ 2 t 2 2! H 2 + γ 3 t 3 3! H 3 + • • •<label>(15)</label></formula><p>The matrix e γtH is named as the diffusion kernel in the sense that the heat diffusion process continues infinitely from the initial heat diffusion.</p><p>γ is an important factor in the diffusion process. If γ is large, the heat will diffuse quickly. If γ is small, the heat will diffuse slowly. When γ → +∞, heat will diffuse immediately, and DiffusionRank becomes into PageRank.</p><p>As in PageRank, there are random relations among vertices. To capture these relations, we use a uniform random relation among different vertices as in PageRank. Let 1 − λ denote the probability that random surfer happens and λ is the probability of following the edges. Based on the above discussion, we can modify DiffusionRank into</p><formula xml:id="formula_178">f (t) = e γtR f (0), R = λH + (1 − λ) 1 N 1<label>(16)</label></formula><p>In application, a computation of e γtR is time consuming. We usually to approximate it to a discrete form</p><formula xml:id="formula_179">f (t) = (I + γ M R) Mt f (0)<label>(17)</label></formula><p>Without loss of generality, we use one unit time for heat diffusion between vertices and their neighbors, we have</p><formula xml:id="formula_180">f (1) = (I + γ M R) M f (0)<label>(18)</label></formula><p>We could iteratively calculate (I+ γ M R) M f (0) by applying the operator (I+ γ M R) to f (0). Therefore, for each iteration, we could diffuse the heat values at each vertices using the following formulation:</p><formula xml:id="formula_181">s = (1 − γ M )s + γ M (λAs + (1 − λ) 1 N 1)<label>(19)</label></formula><p>where M is the number of iterations. As analyzed in <ref type="bibr">[3]</ref>, for a given threshold ǫ, we can compute to get M such that ((I + γ M R) M − e γR )f (0) &lt; ǫ for any f (0) whose sum is one. Similar to <ref type="bibr">[3]</ref>, in this paper we set M = 100 for DiffusionRank.</p><p>Different from FolkRank, in DiffusionRank we set the initial values f (0) for vertices to indicate the preferences. To suggest tags to user u for resource r, we set f (0) = 0, but for f u (0) = 1 and f r (0) = 1. After running DiffusionRank on the tripartite graph, we rank tags according to their ranking scores and the top-ranked tags are suggested to user u for resource r.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Experiments</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1">Data Set</head><p>We use the given BibSonomy data set to validate our methods, it is a snapshot of the BibSonomy system until Jan 1, 2009. The data set contains two parts, bibtex and bookmark. In bibtex, the resources are citation of research papers or books, with title, author and other information. In bookmark, the resources are website URLs with a user-provided short description. Additionally, the contest organizer provide two postcore-2 data sets. In the postcore-2 data sets, the organizer removed all users, tags, and resources which appear in only one post. The process was iterated until convergence and got a core in which each user, tag, and resource occurs in at least two posts. Batagelj et al <ref type="bibr">[9]</ref> provided a detailed explanation of postcore building . The basic statistics of these data sets are lists in Table 1 To validate and tune our methods, we split each of the four dataset into 5 equal-sized subset randomly, and perform 5-fold cross validation on them.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2">Evaluation Metrics</head><p>We use precision, recall and F1 measure as the evaluation metrics. Precision is the number of correct suggested tags multiplied by the total number of tags suggested. Recall is the number of correct suggested tags multiplied by the total number of tags of original post. F1 measure is a geometry mean of precision and recall, F 1 = 2P recsion × Recall/(P recision + Recall). For each post, we only consider the first 5 tags suggested.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3">Content-based Tag Suggestion</head><p>To test the performance of our content-based method, we run 5-fold cross validation using the given training data. Additionally, for each fold, we remove all posts in the postcore set from the test data, since posts in postcore will not appear in the final test data. We remove stopwords, punctuation marks and all words shorter than 2 letters from the data set, and convert all text to lowercase. We remove words, resource IDs and user IDs appear in less than 5 post. We treat bibtex and bookmark separately.</p><p>We use search-based kNN as our baseline method, this is proposed by Mishne <ref type="bibr">[10]</ref> for suggesting tags to blog posts. In our experiment, we index the training data by Lucene<ref type="foot" target="#foot_82">1</ref> indexing package. For a test post, we use T F × IDF to select 10 top words. Then, we use these words to construct a weighted query, and search the training data with it. We take all tags from Lucene returned top-k documents, weight each tag using the corresponding document's relevance score, and sum the weights of duplicated tags. We take the first 5 tags as the suggested tags. In search-based kNN, k is a parameter to tune. After using k = 1, 2, 3, 4, 5, we use k = 1 as the final k, since it has the best F1 measure.</p><p>We list the mean precision, recall and F1 value for bibtex and bookmark data in Table <ref type="table" target="#tab_1">2</ref> and 3 respectively. We experimented with the different combination of methods for weighting features and estimating Θ matrix.</p><p>In the bibtex dataset, FDT(TFITF+MI) has the similar performance as the search-based kNN methods. In the bookmark dataset, FDT(TFITF+MI) has the best performance, which is 3 percentage better than search-based kNN. In the training data, the number of post from each user roughly follows the power law distribution, where most users have less than 100 posts, and the top 4 users have 50% of all posts. If we treat all posts as equal, then the model may bias to the preference of several super users. To know the performance of the methods on super users and common users, we run other two experiments. In the first experiment, we train the model using posts from all users, then check its performance on each of the top n users and all the rest users separately. In </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.4">Graph-based Tag Suggestion</head><p>In experiments, we compare the results of three graph-based methods, most popular tags, FolkRank and DiffusionRank.</p><p>Here we first demonstrate the results using 5-fold cross validation on training dataset. In Table <ref type="table" target="#tab_10">6</ref>, we show the best performance of various methods on bibtex dataset. In this table, we also demonstrate the performance of the content-based method kNN , which achieves the best result when k = 2. For the method of most popular tags, we use "mpt+resource" to indicate most popular tags by resource, and "mpt+mix" to indicate most popular tags by mixing resource and user. For FolkRank, the best result is achieved when damping factor λ = 0.01 with 100 iterations. DiffusionRank obtains the best result when damping factor λ = 0.85, maximum number of iterations max i t = 10 and diffusion factor γ = 0.1. From the table, we can see that most popular tags by mix achieves the best F 1measure, which has the largest precision. While for DiffusionRank, it achieves the best recall. In Table <ref type="table" target="#tab_11">7</ref>, we show the best performance of various methods on bookmark dataset. kNN achieves the best performance when k = 2. For FolkRank, the best result is achieved when damping factor λ = 0.0001 with 10 iterations. DiffusionRank obtains the best result when damping factor λ = 0.85, maximum number of iterations max i t = 10 and diffusion factor γ = 0.01. Furthermore, we also restrict the scores of suggested tags should be no less than 1/5 of score of first-ranked tags. From the table, we can see that DiffusionRank achieves the best F 1 -measure, which has the largest precision. From the above two tables, we find that on the bibtex dataset the method of most popular tags by mix is the best, and on bookmark dataset DiffusionRank achieves the best result. Therefore, for task 2 of rsdc'09, we use the two methods to train ranking models separately on bibtex and bookmark. Using the original result and evaluation program provided by the challenge organizer, we obtain the evaluation results on test dataset, as shown in Table <ref type="table" target="#tab_19">8</ref>  Besides the above analysis, we want to investigate the performance of FolkRank and DiffusionRank as their parameters change.</p><p>In Table <ref type="table" target="#tab_20">9</ref> and 10, we demonstrate the performance of FolkRank on bibtex training dataset and bookmark training dataset as its parameters, the damping factor λ and maximum number of iterations (denoted as "max-it" in tables) change. From the both tables, we find the performance of FolkRank improves as damping factor shrinks, which indicates the effect of preference values are growing larger. That is to say the generalization of FolkRank by passing values iteratively on graphs may harm the performance. Moreover, it seems that the maximum number of iterations of FolkRank does not effect the results significantly.</p><p>In Table <ref type="table" target="#tab_22">11</ref> and 12, we demonstrate the performance of DiffusionRank on bibtex training dataset and bookmark training dataset as its parameters, the diffusion factor γ and maximum number of iterations (denoted as "max-it" in tables) change. Here the damping factor λ is set to 0.85. We also find that the performance of DiffusionRank improves as diffusion factor shrinks, which indicates the effect of initial values is growing larger. Similar to FolkRank, the generalization of DiffusionRank by passing values iteratively on graphs may also harm the performance. It is also the same as FolkRank that the maximum number of iterations of DiffusionRank does not effect the results significantly.</p><p>From the experiments on both bibtex and bookmark training datasets, we can see that DiffusionRank always outperforms FolkRank with some specific parameters, which is more significant on bookmark dataset. Although in this dataset, FolkRank does not outperform the method of most popular tags, in <ref type="bibr">[8]</ref> we know that in some datasets, FolkRank outperforms most simple methods including the method of most popular tags. Therefore, more experiments still need to be done to investigate the efficiency of DiffusionRank compared to FolkRank and other graph-based methods for tag suggestion.</p><p>Furthermore, the number of suggested tags should be specified in advance in FolkRank and DiffusionRank. However in some conditions, we do not have to recommend as many tags as specified. For DiffusionRank, we set the maximum number of suggested tags is 5. If we further require the suggested tags should have the ranking values no less than 1/5 of the ranking value of the first-ranked tag, the performance of precision, recall and F 1 -measure will be improved to 0.3772, 0.3266 and 0.3501 on bookmark training dataset. Therefore, we use the altered DiffusionRank for the bookmark test set of task 2 in rsdc'09. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Related Work</head><p>Ohkura et al <ref type="bibr">[11]</ref> proposed a Support Vector Machine-based tag suggestion system. They train a binary classifier for each tag to decide if this tag should be suggested. Katakis et al <ref type="bibr">[12]</ref> use a hierarchical multi-label text classifier to find the proper tags for a document. They cluster all tags using modified k-means, use one classifier to decide which clusters a document belongs to, then use another cluster-specific classifier to decide which tags in the cluster belongs to the document. Mishne <ref type="bibr">[10]</ref> use a search-based nearest neighbor method to suggest tags, where the tags of a new document is collected from the most relevant documents in the training set. Lipczak et al <ref type="bibr">[13]</ref> extract keywords from the title of a document, then filter them with a user's used tags to get the final suggestion. These methods all use the content of a document, we call them content-based methods. Tatu et al <ref type="bibr">[14]</ref> combine tags from similar documents and extracted keywords to provide tag suggestions. They have the best performance in the first ECML/PKDD Discovery Challenge task. Another class of tag suggestion system is based on the links between users, tags and resources, which does not take the content of resources into consideration. Since the method of "most popular tags" also does not consider the content of resources, in this paper we regard it as a member of graph-based tag suggestion approach. Xu et al <ref type="bibr">[15]</ref> use collaborative filtering to suggest tags for URL bookmarks. Jaschke et al <ref type="bibr">[2]</ref> proposed FolkRank, a PageRank-like iterative algorithm to find the most related tags for a resource in its neighbor users and tags. PageRank is originally used for ranking web pages only according to the topology of web graph. However, in PageRank we can set preference values to a subset of pages to make the PageRank values biased to these pages and their neighbors. In fact, FolkRank is used to compute the relatedness between tags and the specific user and resource by setting the given user and resource to high preference values in PageRank.</p><p>Recently, a new graph-based ranking method, DiffusionRank <ref type="bibr">[3]</ref>, is proposed for anti-spam of web pages. DiffusionRank is motivated by the heat diffusion process, which can be used for ranking because the activities flow on the graph can be imagined as heat flow, the edge from a vertex to another can be treated as the pipe of an air-conditioner for heat flow. Based on the property of heat always flow from high to low, the ranking values of DiffusionRank are related to initial values of vertices. Therefore, DiffusionRank provides a more flexible method to rank tags by setting high initial values to the given user and resource. In this paper, we for the first time propose to use DiffusionRank for graph-based tag suggestions.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7">Conclusion</head><p>In this paper, we study the problem of tag suggestion and describe our methods for content-based and graph-based suggestion. For content-based tag suggestion, we propose a new method named Feature-Driven Tagging for fast contentbased tag suggestion. Cross validation on the training data shows that FDT outperforms wildly-used search-based kNN, especially when suggesting tags for long-tail users. For graph-based tag suggestion, we study most popular tags, FolkRank, and propose a DiffusionRank-based method. Experiments show that on bibtex dataset the method of most popular tags by mixing of user and resource performs best, and on bookmark dataset, DiffusionRank outperforms other methods.</p><p>Work remains to be done. First, currently we use empirical methods to estimate the parameters for FDT, like CC, MI and ITF. We will consider learn a Θ matrix directly by optimizing a tag-related loss function. Second, evaluation using final test data of DC09 shows that the F1 value drops a lot than cross validation on the training data, especially for content-based methods. This suggests we should pay attention to out-of-vocabulary tags. Third, more information should be considered, such like time-stamp, to suggest better tags in real-world situation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Social bookmarking services allow users to share and store references to various types of World Wide Web (WWW) resources. Users can assign tags to these resources, several words best describing the resource content and his or her opinion. To assist the process of assigning tags, some services would provide recommendations to users as references. In Tatu et al. <ref type="bibr">[5]</ref> work, they mentioned that the average number of tags in RSDC'08 bookmarking data is two to three. Thus, it is not an easy task to provide reasonable tag recommendations for the resource with only two to three related tags on average. Tag recommendation is a challenge task in ECML PKDD 2009 where participants should provide either content-based or graph-based methods to help users to assign tags. This work shows some results that aim to this challenge.</p><p>The challenge provides description of the resources and posts of the tag. Description contains some basic information about the resources and post is the tuple of user, tag and resource. In the challenge, there are two types of resources, normal web pages, named as bookmark, and research publications, named as bibtex, with different schemas of descriptions. A post records the resource and the tags assigned to it by a particular user. The task is to provide new tags to new resources with high F-Measure performance on the top five recommendations. The difficulties of this challenge fall in:</p><p>-How to take advantage of the record content itself, while the description is very limited? For example, bookmark is only described with the title of the web page and a short summary while bibtex is usually described with title, publication name, and authors of the paper.</p><p>-How to utilize history information to recommend tags which do not appear in the page content? Though we can use keywords to help find possible tags, tags are not just keywords. Tags could be user's opinion about the page, the category of the page, so on and so forth. This kind of tag might be tracked by using history information. -How to choose the most appropriate two to three tags among the potential pool? By analyzing the page content and history information, we might have a pool which contains the reasonable tag recommendations. Yet we cannot recommend all those to the user. Instead of that, only two to three tags need to be extracted from that pool.</p><p>In order to solve the above problems, we propose tag recommendation using both keywords in the content and association rules from history records. After we end with a pool which contains potential appropriate tags, we introduce a method, named common and combine, to extract the most probable ones to recommend. Our evaluation showed that integrating association rules can give better F-Measure performance than simply using keywords.</p><p>Besides using association rules, some history information will be used more directly, if the resource has been tagged before or the target user tagged other documents before. These history records would greatly improve recommendation performance.</p><p>In this paper, we tuned some parameters in our recommendation system to generate the best F-Measure performance while recommending at most five tags.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Related Work</head><p>Lipczak <ref type="bibr">[3]</ref> proposed a recommendation system mainly based on individual posts and the title of the resource. The key conclusion of their experiments is that, they should not only rely on tags previously attached when making recommendations. Sparsity of data and individuality of users greatly reduce the usefulness of previous tuple data. Looking for potential tags they should focus on the direct surrounding of the post, suggesting a graph-based method. Tatu et al. <ref type="bibr">[5]</ref> proposed a recommendation system that takes advantage of textual content and semantic features to generate tag recommendations. Their system outperformed other systems in last year's challenge. Katakis et al. <ref type="bibr">[2]</ref> proposed a multilabel text classification recommendation system that used titles, abstracts and existing users to train a tag classifier.</p><p>In addition, Heymann et al. <ref type="bibr">[1]</ref> demonstrated that "Page text was strictly more informative than anchor text which was strictly more informative than surrounding hosts", which suggests that we do not have to crawl other information besides page content. They also showed that the use of association rules can help to find recommendations with high precision.</p><p>3 Dataset Analysis and Processing</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Dataset from the Contest</head><p>Three table files were provided by the contest, including bookmark, bibtex and tas. The bookmark file contains information for bookmark data such as contentID, url, url-hash, description and creation date. The bibtex file contains information for bibtex data such as contentID, and all other related publication information. The tas file contains information for (user, tag, resource) tuple, as well as the creation date. The detailed characteristics for these files could be found in Table <ref type="table" target="#tab_0">1</ref>. In this work, all contents were transformed into lower case since the evaluation process of this contest ignores case. In the mean time, we filtered the latex format when we exported bibtex data from the database. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Building Experiment Collection</head><p>We considered and tried merging duplicate records together in training process yet found it did not help much. Thus we kept the duplicate records when building our experiment collections. Since our proposed tag recommendation approach does not involve a training process, we did not separate the dataset into training one and testing one at first. We evaluated our recommendation system on all documents in the given dataset. Based on the type of documents, there are three different collections in our dataset:</p><p>bookmark collection from dataset provided We created a collection bookmark more to contain all bookmark information which were provided by the contest training dataset. Every document in the collection corresponds to a unique contentID in bookmark file. It contains all information for that record, including description and extended description. There are 263,004 documents in this collection.</p><p>During the experiment, we crawled the external webpage for every contentID. Yet the performance showed that the external webpage are not as useful as the simple description provided by the contest. Regardless of performance, it also cost too much time, which is not realistic for online tag recommending. In addition, an external webpage usually contains too many terms, which makes it even harder to extract two to three appropriate terms to recommend as tags.</p><p>bibtex collection from dataset provided We created a collection bibtex original to contain all bibtex information which were provided by the original dataset. Every document in the collection corresponds to a unique contentID in bibtex file. It contains all information for that record, including all attributes in Table <ref type="table" target="#tab_0">1</ref> except simhash0, simhash1 and simhash2. There are 158,924 documents docs in this collection. bibtex collection from external resources If the url of a bibtex record points to some external websites such as portal.acm.org and citeseer, we crawled that webpage and extracted useful information for this record. All these documents are stored in another collection. Similarly, every document in the collection corresponds to a unique contentID in bibtex file. There are 3,011 documents in this collection bibtex parsed.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Keyword-AssocRule Recommendation</head><p>We consider the tag recommendation problem as to find the most probable terms that would be chosen by users. In this paper, P (X) indicates the probability of term X to be assigned to the document as tag. For every document, the term with high P (X) has the priority to be recommended.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Keyword Extraction</head><p>In this step, our assumption is that the more important this term in the document, the more probable for this term to be chosen as tag.</p><p>We used two term weighting functions, TF-IDF and Okapi BM25 <ref type="bibr">[4]</ref> to extract "keywords" from resources. In a single collection, we calculated TF-IDF and BM25 value for every term in every document.</p><p>For TF-IDF, the weighting function is defined as follows:</p><formula xml:id="formula_182">T F − IDF = T F t,d × IDF t<label>(1)</label></formula><p>where T F t,d is the term frequency that equal to the number of occurrences of term t in document d. IDF t is inverse document frequency that is defined as:</p><formula xml:id="formula_183">IDF t = log N df t<label>(2)</label></formula><p>where df t is the number of documents in the collection that contain a term t and N is the total number of documents in the corpus. For Okapi BM25, the weighting function is defined as follows:</p><formula xml:id="formula_184">BM 25 = n i=1 IDF (q i ) T F t,d (1 + k 1 ) T F t,d + k 1 (1 − b + b × L d Lave )<label>(3)</label></formula><p>where T F t,d is the frequency of term t in document d and L d and L ave are the length of document d and the average document length for the whole collection. IDF (q i ) here is defined as</p><formula xml:id="formula_185">IDF (q i ) = log N − n(q i ) + 0.5 n(q i ) + 0.5<label>(4)</label></formula><p>The terms in the single document are ranked according to its TF-IDF or BM25 value in decreasing order. A term with high value or high rank is considered to be more important in the document. Thus P k (X) can be calculated by Algorithm 1.</p><p>Algorithm 1 To calculate P k (X), by using results from keyword extraction method for all documents in the collection do rank all terms according to TF-IDF or BM25 value in decreasing order for all term X in the document do P k (X) = 100 − rank(X); {//rank(X) = 1 indicated the top position, 2 indicated the second position, etc. } end for end for As shown in Table <ref type="table" target="#tab_1">2</ref>, TF-IDF performed better than BM25 in tag recommendation process. The following processes in this work were all performed based on results of TF-IDF method.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Using Association Rules</head><p>Recent work by Heymann et al. <ref type="bibr">[1]</ref> showed that using association rules could help to find tag recommendation with high precision. They expanded their recommendation pool in decreasing order of confidence. In this paper, we used  The rules X → Y we constructed all have support &gt; 10, thus at least 10 resources in our training dataset contain both X and Y as tags. As we mentioned before, we did not separate the dataset into training and testing sets. During evaluation, some records might benefit from the rule it contributed at first, yet at least 9 more resources also contributed to the rule. The support limit here is chosen arbitrarily. Two sets of rules are constructed independently, one for bookmark dataset and another one for bibtex dataset. Some sample rules are showed in Table <ref type="table" target="#tab_3">3</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Choosing appropriate recommendations by using association rules</head><p>Here the problem becomes to be:</p><p>If X → Y exists in the association rules, how possible that term Y should be recommended when X is likely to be recommended? Given P (X) and the confidence value P (Y |X), P (Y ) could be calculated according to law of total probability, which is sometimes called as law of alternatives:</p><formula xml:id="formula_186">P (Y ) = n P (Y ∩ X n )<label>(5)</label></formula><p>or </p><formula xml:id="formula_187">P (Y ) = n P (Y | X n )P (X n ) (6) since P (Y ∩ X n ) = P (Y | X n )P (X n ).</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">Combining Keyword Extraction with Association Rules Results</head><p>After P k (X) and P a (X) are calculated for every term in the document, one method, in Algorithm 3, is to linearly combine the two values to calculate the final probability P c (X) for recommending a term X.</p><p>Similarly, term with higher P c (X), i.e., higher rank in Combined results has the priority to be recommended.</p><p>The experiments showed that weight could affect the F-Measure performance and the optimal weight to combine is different for every collection. Figure <ref type="figure" target="#fig_0">1</ref> shows the effect of weight in bibtex parsed collection, where F-Measure reaches the peak during increase of weight from 0.1 to 0.9. This trend is similar in other two collections. Our experiments indicated that the optimal weight to achieve best F-Measure for bibtex parsed, bibtex original, bookmark more is 0.7, 0.5 and 0.5, respectively. The evaluation results with optimal weight for every collection, in this step, is shown in the second column of Table <ref type="table" target="#tab_5">4</ref>. Compared to the TF-IDF results in the first column, it is obvious that the association rules can greatly help to improve the F-Measure performance.</p><p>Another method we found that worked well is common and combine. In common step, if the term in top rank of keyword extraction results do have Assoc(X) &gt; 0, then recommend this term. In combine step, extract terms with Since the evaluation of this contest only cares for the first 5 tags to recommend, we set k = 5. If common-no = 10 and combine-no = 5, the results for all three collections are shown in third column of Table <ref type="table" target="#tab_5">4</ref>.</p><p>Generally speaking, F-Measure increases with the increase of common-no and reaches the peak near common-no = 20. At the same time, it reaches its highest point as combine-no increases, and remains the same level with the further increase of combine-no. Since the total number of tags to recommend is fixed to be 5, the combine step will stop before it reaches the limit of how many tags to check, i.e., combine-no. Thus if combine-no is greater than a certain number, it won't affect the f-measure performance anymore.</p><p>Since the recommendations would be further modified by history results, we set k = 80, common-no=10, and combine-no=80 here.</p><p>If only recommending at most 5 tags, the F-Measure performance of all above methods, including only using TF-IDF, linearly combining results of TF-IDF &amp; association rules, and common &amp; combine the two, are shown in Figure <ref type="figure" target="#fig_4">2</ref>. It is obvious that using association rules can greatly enhance the TF-IDF performance, either by linear combination or common &amp; combine. Common &amp; combine method is slightly better than linearly combining the two.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4">Checking Resource or User Match with History Records</head><p>In this section, historical information is used more directly. We performed 10-fold cross validation to report the performance in this section. Resource match If the bookmark or bibtex in the testing dataset already appeared before in training dataset, regardless of which user assigned the tags, the tags that were assigned before would be directly inserted into our recommendation list for this document. These tags from historical information have higher priority than the tags that were recommended in previous steps.</p><p>User match Suppose the tags that are assigned by users previously in the training dataset, regardless of to which documents, make up the user's tagging vocabulary.</p><p>Our assumption here is that every user prefers to use tags in his/her own tagging vocabulary, as long as the tags are relevant to the document. Thus the tags in the user's tagging vocabulary would be given higher priority. The common and combine algorithm is again applied here. In common step, if the terms with high rank in previous steps do appear in user's tagging vocabulary, then recommend this term. In combine step, extract terms with high ranks in previous steps to recommend. The number of tags to check in the common step is common-no, and the number of tags to extract in the combine step is combine-no. The two parameters, common-no and combine-no, are tuned to achieve the best F-Measure performance when recommending at most 5 tags. common-no is fixed to be 53 in Figure <ref type="figure" target="#fig_10">3</ref>, while combine-no increases from 1 to 5. In that figure, it shows that F-Measure increases and reaches the peak point at combine-no = 1. In Figure <ref type="figure" target="#fig_15">4</ref>, combine-no is fixed to be 1 and common-no increases from 1 to 80. F-Measure increases with the initial increase of common-no and reaches the peak point in the middle. In this work, we set common-no = 53, and combine-no = 1.</p><p>Exact match with same user and same resource In this step, if user has tagged the same document in the training dataset, then the tags he used before for this document would be directly recommended again.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.5">Combining Results in all Collections</head><p>According to the performance of each collection, our priority to combine the results is shown in Table <ref type="table" target="#tab_7">5</ref>. For example, if a record both exists in bibtex parsed and bibtex original, the results for this record are chosen from bibtex parsed instead of bibtex original, since the former one has higher priority.</p><p>If we only consider to combine the common &amp; combine results for all three collections, the best performance is shown in column without checking the history records of Table <ref type="table" target="#tab_10">6</ref>. If step Tags from records that match with same user has lower priority than tags from records that match with same resource, the best result is shown in column resource match higher of Table <ref type="table" target="#tab_10">6</ref>. Otherwise, the best result is shown in column user match higher of Table <ref type="table" target="#tab_10">6</ref>. The results indicate that even for those bookmarks that were tagged by other users before, it is still beneficial to consider the target user's own tagging vocabulary.</p><p>To sum up, the best performance on training dataset is shown in Table <ref type="table" target="#tab_11">7</ref>, including the detailed results only for bookmark and bibtex. or target user appeared before, the history tags would be used as references to recommend, in a more direct way. Our experiments showed that association rules could greatly improve the performance with only keyword extraction method, while history information could further enhance the F-Measure performance of our recommendation system. In the future, other keyword extraction method can be implemented to compare with TF-IDF performance. In addition, graph-based methods could be combined with our recommendation approach to generate more appropriate tag recommendations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Understanding the user: Personomy translation for tag recommendation</head><p>Robert Wetzker 1 , Alan Said 1 , and Carsten Zimmermann 2</p><p>1 Technische Universität Berlin, Germany 2 University of San Diego, USA</p><p>Abstract. This paper describes our approach to the challenge of graphbased tag recommendation in social bookmarking services. Along the ECML PKDD 2009 Discovery Challenge, we design a tag recommender that accurately predicts the tagging behavior of users within the Bibsonomy bookmarking service. We find that the tagging vocabularies among folksonomy users differ radically due to multilingual aspects as well as heterogeneous tagging habits. Our model overcomes the prediction problem resulting from these heterogeneities by translating user vocabularies, so called personomies, to the global folksonomy vocabulary and vice versa. Furthermore we combine our user-centric translation approach with item-centric methods to achieve more accurate solutions. Since our method is purely graph-based, it can also readily be applied to other folksonomies.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Over the last years, social bookmarking services, such as Delicious<ref type="foot" target="#foot_84">3</ref> , Bibsonomy<ref type="foot" target="#foot_85">4</ref> and CiteULike<ref type="foot" target="#foot_86">5</ref> , have grown rapidly in terms of usage and perceived value. One distinguishing feature provided by these services is the concept of tagging -the labeling of content with freely chosen keywords (tags). Tagging enables users to describe and categorize resources in order to organize their bookmark collections and ease later retrieval. Social bookmarking services are therefore the classic example of collaborative tagging communities, so called folksonomies [1] <ref type="bibr">[2]</ref>. The consumer-centric (collaborative) tagging aspect differentiates social bookmarking from other content sharing community services, such as Flickr<ref type="foot" target="#foot_87">6</ref> or YouTube <ref type="foot" target="#foot_88">7</ref> , where tags are generally assigned by the content creator <ref type="bibr">[3]</ref>. Most folksonomy solutions assist users during the bookmarking process by recommending tags. Thus, a user can select recommended tags from various sets in addition to entering tags manually. Despite their positive effect on usability, these recommenders are effective tools to limit tag divergence within folksonomies as they are generally considered to lower the ratio of misspellings and to increase the likelihood of tag reassignments. The design of a folksonomy tag recommender was one of the tasks of the ECML PKDD 2009 Discovery Challenge<ref type="foot" target="#foot_89">8</ref> . In the following sections, we describe our solution to this task as submitted.</p><p>Our approach is based on the observation that the tag vocabularies of users, their personomies, differ within a folksonomy. This heterogeneity is mainly caused by differences in the tags users constantly assign to categorize content and the multilingualism of the user base, as apparent for Bibsonomy. To overcome the problems caused by this heterogeneity, we propose a tagging model that translates the personomy of each user to the folksonomy vocabulary and vice versa. We find that our model is highly accurate as it characterizes an item by its tag spectrum before translating this spectrum to a user's personomy. We then combine the translational model with item-centric tag models to improve performance.</p><p>This paper is structured as follows: The introductory section presents a graph model for the underlying data structure and explains the actual goals of the challenge. This is followed by an analysis of different properties of the Bibsonomy dataset with respect to their impact on tag recommendation. Section 3 introduces and discusses the tag vocabularies found within folksonomies, before we present our recommendation algorithm and evaluation results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.1">Modeling folksonomies</head><p>According to <ref type="bibr">[2]</ref>, a folksonomy can be described as a tuple F := (I, T, U, Y ), where I = {i 1 , . . . , i k }, T = {t 1 , . . . , t l } and U = {u 1 , . . . , u m } are finite sets of items, tags and users, and Y is a ternary relation whose elements are called tag assignments. A tag assignment (TAS) is defined by the authors as relation Y ⊆ U × T × I, so that the tripartite folksonomy hypergraph is given by G = (V, E), where V = I ∪ T ∪ U and E = {(i, t, u)|(i, t, u) ∈ Y }. The set of all bookmarks is then given as BM = {(i, u)|∃t : (i, t, u) ∈ Y }. This graph structure is characteristic for all folksonomies.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.2">The challenge: Graph-based tag recommendations (Task 2)</head><p>The ECML PKDD 2009 Discovery Challenge consists of different tasks related to the problem of tag recommendation. Testbeds for all tasks are different snapshots of the Bibsonomy bookmarking service. Our solution contributes to the task of "Graph-Based Recommendations (Task 2)" as it does not consider the content of the given resources. The recommendation task in this setting resembles the problem of link prediction within G given a user and an item node. The recommender thus needs to estimate P (t|i, u), the probability of observing a given tag when a combination of user and item has been observed.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.3">Related work</head><p>One of the first works on tag recommendation in folksonomies is <ref type="bibr">[4]</ref>, where the authors compare the performance of co-occurrence based tag recommenders and more complex recommenders based on the FolkRank algorithm <ref type="bibr">[2]</ref>. Furthermore they report only minor improvements of the FolkRank method over the cooccurrence approaches. Parts of their analysis is performed on a snapshot of the Bibsonomy dataset.</p><p>Further related work was presented by the participants of last years challenge on tag recommendation. The authors of <ref type="bibr">[5]</ref> enrich tag vocabularies by terms extracted from all bookmarks' meta data, such as user given descriptions or the abstract, author, year etc. information in case of publications. More similar to our work is the approach in <ref type="bibr">[6]</ref>. The team combines the keywords found within a resource's title and the actual tags previously assigned to a resource with the tags from a user's personomy. Each vocabulary is mapped to the global tag vocabulary using co-occurrence tables. This mapping is similar to our translation process. However, the fusion of different sources differs from our method and no user-centric optimization of parameters was reported.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">The dataset</head><p>The Bibsonomy bookmarking service allows its users to bookmark URLs and publications in parallel. This hybrid approach makes Bibsonomy different from other bookmarking communities such as Delicious or CiteULike. For each web bookmark, participants are given the URL, the title and an optional description of the resource as provided by the user during the bookmarking process. Bookmarked publications generally come with information about the title, the authors, the abstract or other common bibliographic attributes. The completeness of this information is not guaranteed, and many attribute fields are left empty. Table <ref type="table" target="#tab_0">1</ref> gives an overview of the node statistics found within the dataset for p-core levels one and two<ref type="foot" target="#foot_90">9</ref> . For the construction of our recommender we ignore the meta-data attached to the bookmarked content and only consider the graph G. Furthermore, we do not distinguish between URLs and publications, but merge both node sets to the item set I. Both decisions result in a loss of information which potentially to occur within this distribution. A recommender that only considers the global tag distribution would assume P (t|u, i) ≈ P (t). We will refer to such a recommender as MostPopular recommender. Item tag distributions. This is the distribution of previously assigned tags for a given item. It was shown that the tag distributions of items converge to a characteristic tag spectrum over time. Furthermore, as reported by <ref type="bibr">[8]</ref> and <ref type="bibr">[9]</ref>, the resulting tag distributions often follow a power law with few tags being assigned very frequently and most tags occurring in the long tail.</p><p>If we neglect the personalization aspect of tagging, we can recommend tags by assuming P (t|i, u) ≈ P (t|i). However, our observation of P (t|i) within the training data may be limited, i.e. information about most tags will be missing. User tag distributions (Personomies). Each user develops his own vocabulary of tags over time called his personomy. Users will generally be interested in reassigning previously used tags as this will simplify content search later on. The interest in tag convergence often results in the frequent assignment of a limited number of category tags, and it was shown that user vocabularies develop power law characteristics over time <ref type="bibr">[10]</ref>. A personomy based recommender would assume P (t|i, u) ≈ P (t|u). Once again, the distribution P (t|u) estimated over the training data is likely to miss a variety of tags especially for users with few bookmarks.</p><p>The authors of <ref type="bibr">[4]</ref> report that a tag model which combines user and item tag distributions into a unified distribution achieves sufficient recommendation accuracy. We will consider a hybrid recommder with P (t|i, u) ≈ αP (t|i) + (1 − α)P (t|u) as an additional baseline approach during our evaluations (MostPopu-lar2d ).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Our approach</head><p>The design of our tag recommender is based on two intuitive assumptions:</p><p>1. Tags are personalized. Different users will assign different tags to the same item. This effect cannot only be explained by statistical variance. Instead, we find users developing their own category tags over time. One of the implications for the tag recommendation task is the problem of recommending the right version of a tag, especially in cases where synonymous tags exist. This includes different spellings, such as "web20" versus "web2.0". Even though semantically equal, these will be different for a user who assigns tags for content categorization. Furthermore, especially in multilingual folksonomies, we find that users often assign keywords from their mother tongue. This is of particular importance for Bibsonomy, where many users seem to come from Germany, with the effect that the tag distribution is a mixture of German and English words. Whereas some users tagged a site as "searchengine" related we also find the German translation ("suchmaschine") among the frequent tags.</p><p>2. Tags describe items. The authors of <ref type="bibr">[11]</ref> report that the vast majority of assigned tags on Delicious identify the topic, the type or the owner of a URL. We can therefore assume that users will assign personomy tags depending on the item. This assumption is a simplification as it excludes tags, such as "toread" or "self", which actually refer to the user item relationship instead of the item itself. However, we believe that these tags cannot be easily predicted based on the given training data but require deeper knowledge about the user. Luckily, as reported by <ref type="bibr">[12]</ref> for Delicious, usage context and self reference tags are relatively scarce compared to descriptive tags.</p><p>These assumptions directly influenced the basic design decisions for our recommender which suggests tags from a user's personomy with respect to the community opinion about the underlying item.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Translating personomies</head><p>We assume that each user has a distinctive vocabulary of tags. These tags can be translated to the community tag vocabulary by looking at co-occurrences within the shared item space. We are thus interested in the probability P (t u |t, u) that a user will assign a tag t u from his personomy as next tag given that the item was tagged as t by another user. Based on previous knowledge about the users tagging behavior we can estimate P (t u |t, u) as</p><formula xml:id="formula_188">P (t u |t, u) = i∈Iu P (t u |i, u)P (i|t),<label>(1)</label></formula><p>where I u is the set of items previously bookmarked by the user. Based on P (t u |t, u) we can now translate the global folksonomy language to the personomy of a user.</p><p>For the recommendation task we are interested in the probability P (t u |i, u) for previously unseen user item combinations. For a new item with an observed tag vocabulary of P (t|i), the probability that a user will assign one of his tags next is given as P (t u |i, u) = t∈Ti P (t u |t, u)P (t|i).</p><p>(2) Note that t u ∈Tu P (t u |i, u) = 1 is only true if all tags in T i have a mapping in T u which is rather unlikely. Instead we expect t u ∈Tu P (t u |i, u) to decrease the more the given item deviates from the items previously bookmarked by the user.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">The tag recommender</head><p>The tag recommender we propose, selects tags coming from three sources: the personomy of a user where tags are weighted by P (t u |i, u), the item vocabulary (P (t|i)) and the global vocabulary (P (t)). Including the item vocabulary is important, as many tags may be item specific and are thus unlikely contained within the personomy. We also include the global tag distribution P (t) for cases where little is known about user and item alike. We assume a weighted linear combination of sources and estimate P (t|i, u) as P (t|i, u) ≈ α u P (t u |i, u) + α i P (t|i) + (1 − α u − α i )P (t),</p><p>with 0 ≤ α u , α i , α u + α i ≤ 1. We then recommend the N tags with highest probability P (t|i, u), where N is a parameter that needs to be optimized together with α u and α i . Optimization can take place on a global or a user-centric scale.</p><p>Global optimization. For the globally optimized model, we assume equal parameter settings for all users. As the Bibsonomy dataset is rather small, we can use a brute-force approach to find the combination of N , α u and α i that maximizes the F-measure. We do so by performing a 10-fold cross-validation on the training data. We call this user-centric tag recommender with global optimization U C G .</p><p>User-centric optimization. As users are heterogeneous, it is not intuitive to assume shared parameter preferences. Instead, it seems straightforward to optimize parameters for each user separately. Once again, we do so by performing a cross-validation on the training data. We then use a brute-force approach to find the combination of N , α u and α i that maximizes the F-measure of each user averaged over all folds. We will refer to the user-centric tag recommender with local optimization as U C L .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Evaluation</head><p>We trained the models of all recommender types on the 2-core version of the dataset. The parameters of the MostPopular2d, U C G and the U C L models were fine-tuned in a 10-fold cross-validation as described above. For the MostPopu-lar2d recommender we found an α value of 0.5 to perform best. For the usercentric tag recommender with global optimization the maximal F 1 measure was achieved when setting α u = 0.6 and α i = 0.4. The weight of the global tag distribution thus resulted 0 which means that including the global vocabulary did not yield performance gains. For the MostPopular2d as well as the U C G recommender the best number of tags to recommend was 5. Evaluating the accuracy of all recommender types during the cross validation, we found the user-centric tag recommender with local optimization (U C L ) to constantly outperform all other versions. We therefore submitted the predicted tags of the U C L approach as our solution to the challenge. The released test dataset consists of 778 bookmarks from 136 users linking to 667 items. Table <ref type="table" target="#tab_1">2</ref> presents the achieved F 1 measures on the first five ranks for the various recommender types. The U C L recommender we submitted achieved a Table <ref type="table" target="#tab_1">2</ref>. Performance of various recommender types on the test data. Underlined values represent the configuration that performed best during training. The submitted recommender (U CL) achieved an F1 measure of 0.314. The best F1 measure could have been achieved with a U CG recommender always suggesting 3 tags (bold).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Recommender</head><p>F1@1 F1@2 F1@3 F1@4 F1@5 M ostP opular 0.021 0.038 0.051 0.051 0.059 M ostP opular2dα=0.5 0.229 0.286 0.306 0.313 0.310 U CG,α u =0.6,α i =0. <ref type="bibr">4</ref> 0.246 0.326 0.335 0.334 0.330 U CL 0.230 0.294 0.306 0.311 0.314 performance of 0.314. This result is somewhat disappointing as it is only slightly above the result of the simpler M ostP opular2d recommender. However, we find that the approach of vocabulary translation is generally superior as the results of the U C G recommender are significantly better. We observe similar performance patterns when looking at the precision/recall curves plotted in Figure <ref type="figure" target="#fig_4">2</ref>.</p><p>Investigating the reasons for the weak performance of the U C L recommender, we find that the user distribution of the test set deviates from the trained one as shown in Figure <ref type="figure" target="#fig_10">3</ref>. This deviation is likely to have a negative impact on the prediction quality as parameters have been tuned in expectation of a user distribution similar to the one of the training set. However, this problem is not  unique to the U C L recommender but is expected to have a negative impact on the performance of all recommender types. Instead, we believe that the weak performance of the U C L recommender is caused by an inadequate parameter tuning for users less present in the training data but frequent in the test set. Tuning α u , α i and N for these users often results in bad estimates due to missing data. Whereas the implications of these shortcomings are rather minor when users are distributed as in the training data, they seem to become major for the test distribution.</p><p>The fact that the test set is dominated by users with rather small training vocabularies is also reflected by the performance of the M ostP opular2d recommenders with α set to 0 and 1 as shown in Figure <ref type="figure" target="#fig_4">2</ref>. Here, we find that a recommender which only suggest tags from a user's personomy (α = 1) performs very bad, whereas an item based recommender (α = 0) achieves nearly as good results as the mixture model α = 0.5. This implies, that most tags of the test data are not present within a user's personomy at training time or, less likely, that the tagging behavior of users drastically changed in the test phase.</p><p>The inadequate modeling of infrequent users (and items) is an expected shortcoming of a purely graph-based recommendation approach. This is especially true for our personomy translation approach which requires the tags of the given item to have a mapping within the conditional distribution P (t u |t, u) (see equation 2). Incorporating the provided item meta-data may be a promising alternative to improve accuracy in scenarios where little is known about users and items from a graph perspective.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Conclusions</head><p>In this paper, we presented a novel approach to the challenge of graph-based tag recommendation in folksonomies. Building on the assumption that all users of a folksonomy have their own tag vocabulary, our approach translates the personomies of users to the global folksonomy vocabulary. Evaluation results show that this translation helps to significantly improve tag prediction performance. Furthermore, we fine-tuned our model by estimating parameters on a per user basis. Even though this user-centric approach performed rather disappointing during the challenge, we believe that user-level optimization will be essential for the success of future (tag) recommenders.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>With the event of social resource sharing tools, such as BibSonomy<ref type="foot" target="#foot_92">1</ref> , Flickr<ref type="foot" target="#foot_93">2</ref> , del.icio.us <ref type="foot" target="#foot_94">3</ref> , tagging has become a popular way to organize the information and help users to find other users with similar interests and useful information within a given category. Tags posted by a user are not only relevant to the content of the bookmark but also to the certain user. According to <ref type="bibr">[3]</ref>, the collection of a user's tag assignments is his/her personomy, and folksonomy consists of collections of personomies. From the available training data, we can find that some tags might just be words extracted from the title, some tags might be the concept or main topic of the resource, and other might be very specific to a user.(see Table <ref type="table" target="#tab_0">1</ref>). The last three lines of the table show that the user 293 post tags like swss0603, swss0609, swss0602, which are very specific to the user. Since the test data for task one contains posts whose user, resource or tags are not in the training data, some traditional collaborative recommendation systems might not perform well. It is because most of the collaborative recommendation systems cannot recommend tags which are not in the tag set of the training data. This paper presents our tag recommendation system, which is a combination of two methods: simple Language model and an adaption of topic model according to <ref type="bibr">[7]</ref>. The first method can extract some keywords from the description and other information of the post and they constitute a candidate set. Then we use the relevance of a word and a document to score the words in the candidate set and recommend the words with highest scores. The second method uses an ACT model <ref type="bibr">[7]</ref>, and it can get some conceptual or topic knowledge of the post. Given a test post, the model can score all the tags which have been posted previously and recommend the tags with highest scores.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>USER</head><p>These two methods focus on two different aspects. The first method will probably recommend some keywords extracted from the title while the second method uses a probabilistic latent semantic method to recommend tags which are similar to the post in terms of conceptual knowledge. Comparing these two methods, we can find that the tags recommended are always different. Consequently, the combination is a intuitive better way.</p><p>This paper is organized as follows: Section 2 reviews recent development in the area of social bookmark tag recommendation systems. Section 3 describes our proposed system and the combination method in details. In section 4, we present and evaluate our experimental results on the test data of ECML PKDD challenge and conclude the results in section 5.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>2</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Related work</head><p>The recent rise of Web 2.0 technologies has aroused the interest of many researchers to the tag recommendation system. Some approaches are based on collaborative information. For example, AutoTag <ref type="bibr">[9]</ref> and TagAssist <ref type="bibr">[8]</ref> use some information retrieval skills to recommend tags for weblog posts. They recommend tags based on the tags posted to the similar weblogs and they cannot recommend new tags which are not in the training file. FolkRank <ref type="bibr">[4,</ref><ref type="bibr">5]</ref>, which is an adaption of the famous PageRank algorithm, is a graph-based recommendation system. Also, it cannot recommend new tags not in the training file. The experimental results of FolkRank in <ref type="bibr">[5]</ref> reveal that the FolkRank can outperform other collaborative methods. But to some extents FolkRank relies on a dense core of training file and it might not be fit to our task.</p><p>All the methods mentioned above are based on collaborative information and similarity between users and resources. However, in the cases when there are many new users and resources in the test data (our task one), those methods cannot perform well. In the RSDC '08 challenge, the participants <ref type="bibr">[1,</ref><ref type="bibr">2]</ref> who use methods based on words extracted from the title or semantic knowledge and user's personomy can outperform other methods. Consequently, we propose our tag recommendation system mainly based on the contents.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>3</head><p>Our Tag Recommendation System</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Notations</head><p>First, we define notations used in this paper. We group the data in bookmark by its url_hash and data in bibtex by its simhash1. If some posts in bookmark or bibtex file have the same url_hash or simhash1, they are mapped to one resource r. In bookmark, we extract description, extended description while in bibtex, we extract journal, booktitle, description, bibtexAbstract, title and author. We define these information as the description of resource r. For each resource r and each user u who has posted tags to resource r, assuming that its description contains a vector 𝐰 d of N d words; a vector 𝐭 d of T d tags posted to this resource r by the user u d .Then the training dataset can be represented as </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>D w t u w t u </head><p>Table <ref type="table" target="#tab_1">2</ref> summarizes the notations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Language Model</head><p>Language model is widely used in natural language processing applications such as speech recognition, machine translation and information retrieval. In our model, first we pick some words to form a candidate set of recommended tags and then score all the words in the candidate set and recommend words with highest scores for our tag We extract useful words from the description of the active resource r * in the test data. Then we remove all the characters which are neither numbers nor letters and get rid of the stop words in the English dictionary. The rest words form the part of the candidate set C 1. For each t 1 ∈ C 1 , we have the following generative probability: <ref type="bibr">(1)</ref> where N d is the number of word tokens in the description d of r * , tf(t 1 ,d) is the word frequency(i.e., occurring number) of word t 1 in the description of d of r * , N D′ is the number of word tokens in the entire test dataset, and tf(t 1 ,D') is the word frequency of word t 1 in the collection D'. λ is the Dirichlet smoothing factor and is commonly set according to the average document length, i.e. N D ′ /|D′| in our cases.</p><p>As for C 2, in order to get more information about the new resource, we take the similarity between resources into consideration and add tags previously posted to the similar resource into C 2 . The similarity of resource is determined by the url of the resource. Each url can be split into several sections, for example, 'http://www.kde.cs.uni-kassel.de/ws/dc09' will be split into three sections: 'www.kde.cs.uni-kassel.de', 'ws' and 'dc09'. The similarity between r 1 and r 2 is defined as follows, sim(r 1, r 2 ) = 2^(number of the identical sections of url 1 and url 2maximum number of sections of url 1 and url 2 ). For each resource r, we will choose three most similar urls to the url of r and their corresponding resources form the Because the user is a new one, so that we have nothing about his/her history of posting tags in our training data, in that case equation ( <ref type="formula" target="#formula_5">4</ref>) is reasonable. Secondly, not all the words in the given resource are in our training data, so when calculating equation ( <ref type="formula" target="#formula_3">3</ref>), we will ignore the word which has not appeared in the training data. The set of recommended tags for a given user u' and a given resource r' will be: T u′, r′ ≔ argmax t∈T n P(t|u′, r′) where n is the number of recommended tags, T is the collection of tags posted in the training file.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Symbol</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4">Combination</head><p>We have proposed two different methods to recommend tags, model one focuses on the useful words extracted from the description or title of the resource while the second model focuses on the conceptual knowledge and probabilistic relations among tags, resource and users. We are interested in the following problem: Can we combine these two models to perform a better result for tag recommendation?</p><p>Algorithm 1: The combined tag recommendation system</p><p>We have tried some different approaches to combine these two models. A simple method is to combine the scores of these two models and recommend tags with highest scores after combination (Algorithm 1). We can make use of the two scores calculated in the two approaches and there are two things worthy to be noted here: 2) due to the different distribution of scores, we need to normalize the two scores before combination. In order to solve the first problem, we consider all the tags t ∈ C ∪ T where C is the candidate used in the model one. In terms of normalization, we make the score1 ∞ = score2 ∞ and then add these two score, if a tag t is in the candidate set C but not in the T, the score2[t] = 0 and if a tag t is in T but not in C, then the score1[t] =0.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>4</head><p>Experimental Results</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Dataset</head><p>We evaluate our experimental results using the evaluation methods provided by the organizers of ECML PKDD discovery challenge 2009. The training set and the test set are strictly divided and we use the cleaned dump as our training set for our tag recommendation system.</p><p>Here are some statistical information about training data and test data: There are 1,401,104 tag assignments. 263,004 bookmarks are posted and among which there are 235,328 different url resources while 158,924 bibtex are posted and among which there are 143,050 different publications. From this, we can see that many resources appear just once in the training file. There are 3,617 users and 93,756 tags in all in the training file. The average number of tags posted to bookmark is 3.48 and the average number of tags posted to bibtex is 3.05.</p><p>In the test data, there are 43,002 tag assignments, 16,898 posted bookmarks and 26,104 posted bibtex. Among all the posts, there are only 1,693 bookmark resources and 2,239 bibtex resources which are in the training file. The average number of tags posted to bookmark is 3.81 and the average number of tags posted to bibtex is 3.82.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Data Preprocessing</head><p>The training data is provided by the organizers of the ECML PKDD, we establish three tables, bookmark, bibtex and tas in our MySQL database. In order to get the similarity between resources, we need to preprocess the url field. For each url in the bookmark, we eliminate the prefix such as 'http://', 'https://' and 'ftp://'. Then we split the url by the character '/'. For example, a url 'http://www.kde.cs.unikassel.de/ws/dc09' will be split into 'www.kde.cs.uni-kassel.de', 'ws' and 'dc09'. As we mentioned above, we define some information extracted from the table as the description of a resource r. We eliminate the stop words in the English dictionary and stem the words, for example, 'knowledge' will be stemmed to 'knowledg' and both 'biology' and 'biologist' will be stemmed to 'biologi'.</p><p>performance of bookmark is a little bit better than the bibtex and the reason might be that description in bibtex has some information which is irrelevant to the main topic of the publication.  <ref type="table" target="#tab_5">4</ref>: performance of ACT model on the test data, the numbers are shown in the following format: recall/precision/f-measure</p><p>The result after combination is shown in Fig. <ref type="figure" target="#fig_0">1</ref>, together with the results of the previous two methods.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Conclusions</head><p>In this paper, we describe our tag recommendation system for the first task in the ECML PKDD Challenge 2009. We exploit two different models to recommend tags. The experimental results show that the Language model works much better than the ACT model and the combination of these two methods can improve the results. We need to further analyze the results to see why ACT has a poor result for the test data. Also, we can try to change the scoring scheme or expand the candidate set in the language model. Future work also includes some new methods of combination.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Tagging is very useful for users to figure out other users with similar interests within a given category. Users with similar interests might post similar tags and similar resources might have similar tags posted to them. Collaborative filtering is widely used in automatic prediction system. The idea behind it is very simple: those who agreed in the past tend to agree again in the future. Traditional collaborative filtering systems have two steps. The first step is to look for users who share the same rating patterns with the active user whom the prediction is for. Then, the systems will use the ratings from those like-minded users found in the first step to calculate a prediction for the active user. Since all the tags, users and resources in the test data are also in the training file, we can make use of the history of users' tag, also called personomy <ref type="bibr">[3]</ref> and tags previously posted to the resource to recommend tags for a active post. This paper presents our proposed tag recommendation system, which is a combination of two methods: one is an adaption of item-based collaborative filtering, the other is FolkRank according to <ref type="bibr">[4,</ref><ref type="bibr">5]</ref>. As we mentioned above, collaborative filtering performs well for automatic prediction. However, current widely used collaborative filtering systems are for predicting the ratings of some products or recommend some products to users. For example, the famous websites, Amazon.com 1 , Last.fm<ref type="foot" target="#foot_96">2</ref> , eBay<ref type="foot" target="#foot_97">3</ref> apply this method to their recommendation systems. Our first method considers the tags previously posted to the resource and users' similarities to recommend tags. The second method is an application of the FolkRank algorithm in <ref type="bibr">[4,</ref><ref type="bibr">5]</ref>.</p><p>These two methods have some common features. They both use the history of the user and tags previously posted to resource for recommendation. They are both suitable to the case that test data are in the training data. Both of them do not need to establish models in advance. But they are different to some extents. The first method just considers tags in the candidate set while the FolkRank will consider all the tags in the training data. Moreover, the first method focuses more on collaborative information while the second focuses on the graph information.</p><p>This paper is organized as follows: Section 2 introduces recent trends in the area of social bookmark tag recommendation systems. Section 3 describes our proposed system and the combination method in details. In Section 4, we present and evaluate our experimental results on the test data of ECML PKDD challenge 2009 and make some conclusions in Section 5.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Related work</head><p>Some researchers have already used some approaches based on collaborative information for tag recommendation systems. For example, AutoTag <ref type="bibr">[7]</ref> and TagAssist <ref type="bibr">[6]</ref> make use of information retrieval skills to recommend tags for weblog posts. They recommend tags based on the tags posted to the similar weblogs. Our first method is similar to these two approaches. FolkRank in <ref type="bibr">[4,</ref><ref type="bibr">5]</ref> is a topic-specific ranking in folksonomies. The key idea of FolkRank algorithm is that a resource which is tagged with important tags by important users becomes important itself. In <ref type="bibr">[5]</ref>, the author compared the performance of some baseline methods and his FolkRank algorithm, and found that FolkRank outperformed other methods. His experimental results relied on a dense core of the training file and considering that our training data is a post-core two dataset, we decide to refer to this algorithm in our proposed tag recommendation system.</p><p>In the RSDC '08 challenge, the participants <ref type="bibr">[1,</ref><ref type="bibr">2]</ref> who make use of resource's similarities and users' personomy outperformed other approaches. Consequently, we consider using the collaborative information of resource's similarities and users' personomy in our tag recommendation system.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>3</head><p>Our Tag Recommendation</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Notations</head><p>First, we define notations used in this paper. We group the data in bookmark by its url_hash and data in bibtex by its simhash1. If some posts in bookmark of bibtex file have the same url_hash or simhash1, they are mapped to one resource r. For each resource r d , assuming a vector 𝐭 d of T d tags posted to this resource r d by the user u d .</p><p>Then the training dataset can be represented as the candidate set of tags to be recommended for a given user u and a given resource r T (u,r) the set of tags that will be recommended for a given user u and a given resource r n(t, r) the number of times that the tag t has been posted to the resource r in the training dataset Table1: Notations</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Collaborative Filtering method</head><p>Our proposed collaborative filtering method for tag recommendation has two steps. First of all, for a given resource r and a given user u in the test dataset, we make use of the tags previously posted to the resource r in the training dataset and define them as the candidate set: (1 )</p><formula xml:id="formula_190">w dAw d p       (<label>3</label></formula><formula xml:id="formula_191">)</formula><p>where A is the adjacency matrix of G, p is the random surfer component, and d ∈ [0,1] is a constant which controls the influence of the random surfer. Usually, p is set to the vector where all values equal to 1. But in order to recommend tags relevant to certain user and certain resource, we can change the p to express user preferences. In our tag recommendation system, each user, tag, and resource get a preference weight of 1 but the active user and resource for recommendation get a preference of 1+|U| and 1+|R| respectively.</p><p>The FolkRank algorithm has a differential approach to see the ranking around the topics defined in the preference vector. This approach is to compare the rankings with and without the preference vector p . Assuming that 𝐰 0 is the ranking after iteration with d = 1 while 𝐰 1 is the ranking after iteration with d =0.625, then the final weight will be 𝐰 = 𝐰 𝟏 − 𝐰 0 . Details can be found in Algorithm 1.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Experimental Result</head><p>As performance measures we use precision, recall and f-measure. For a given user u and a given resource r, the true tags are defined as TAG(u,r), then the precision, recall and f-measure of the recommended tags T (u, r) are defined as follows: </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.1">Performance of Collaborative Filtering method</head><p>In table 4, we show the performance of collaborative filtering method on the test data provided by the organizers of ECML PKDD challenge 2009. From the table, we can see that this method has a highest f-measure of 30.002% when the number of recommended tags is 5.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.2">Performance of FolkRank method</head><p>In table 4, we show the performance of FolkRank algorithm on the test data. From the table, we can find that the first method performs a little bit better than FolkRank and FolkRank has a highest f-measure of 28.837% when the number of recommended tags is 4.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Ruser (top left), Rres (bottom left) and R res user (right) of a given test post (nodes in grey)</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>φ</head><label></label><figDesc>res-tag := (|Y ∩ (U × {res(x)} × {t})|) t∈T 3. Similar to 2, but x and x are represented as resource-user profile vectors where each component corresponds to the count of co-occurrences between resources and users: φ res-user := (|Y ∩ ({u} × {res(x)} × T )|) u∈U 4. The same as in 1, but the node similarity is computed w.r.t. to user-resource profile vectors: φ user-res := (|Y ∩ ({user(x)} × {r} × T )|) r∈R</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Fig. 2 .Fig. 3 .</head><label>23</label><figDesc>Fig. 2. Parameter search of WA*-Full in a holdout set. Best c value found equals 3.5</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 1 .</head><label>1</label><figDesc>Figure 1. Tag recommendation process.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 2 .</head><label>2</label><figDesc>Figure 2. Filtered tag co-occurrence graph associated to the example input bookmark. Edge weights and non-connected vertices are not shown. Two main clusters can be identified in the graph, which correspond to two research areas related to the bookmarked document: recommender systems, and semantic web technologies.</figDesc><graphic coords="27,129.14,238.74,338.40,396.36" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Example of ranking SVM model</figDesc><graphic coords="38,214.24,394.52,170.00,113.30" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head>Insider 3 | 7 | 9 |LinuxQuestions.org 11 |</head><label>37911</label><figDesc>programming | xwindows | google | gui | software | microsoft | portables | os | storage | linuxbusiness | security | gnu | education | caldera | portables Linux.com :: Features 8 URL broken HowtoForge -Linux Howtos and Tutorials Ubuntu | Debian | Ubuntu | Desktop | Debian | Lighttpd | Ubuntu | Desktop | Virtualization | Ubuntu | Desktop | Security | Ubuntu | CentOS | Samba | Ubuntu | Desktop | Linux | Ubuntu | Security | Ubuntu | Desktop | Fedora | Security Linux and Open Source -RSS Feeds 10 No Categories Linux -Newbie | Linux -Newbie | Linux -Newbie | Linux -Software | Programming | Red Hat | Linux -General | Linux -General | Linux -Laptop and Netbook | Puppy | Ubuntu | Linux -Desktop | Ubuntu | Ubuntu | Linux -Security LXer Linux News 12 | linux | linux | linux | linux | linux | linux | linux | linux | linux | linux | linux | linux | linux | linux | linux | linux | linux | linux | linux</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_7"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Values of precision, recall and F1 measures for different levels of overlap threshold.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_8"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Informational channels of a folksonomy.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_9"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. Evaluation of recommendation techniques: recall vs. precision.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_10"><head>Fig. 3 .</head><label>3</label><figDesc>Fig. 3. Evaluation of recommendation techniques: F1-measure.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_11"><head></head><label></label><figDesc>0 and the other where − → W • − → x i + b &lt; 0. Every vector for which − → W • − → x i + b = 0 lies on hyperplane h. The objective of each linear classifier is to define such h : − → W , b . Different linear classifiers have different ways to define model vector − → W and bias b.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_12"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. A simple Centroid Classifier in 2 dimensions. The positive class C+ is linearly separable from the negative one C−.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_13"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. No perfect separating hyperplane exists for this Centroid Classifier. Dark regions are misclassified.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_14"><head>Fig. 3 .</head><label>3</label><figDesc>Fig. 3. The proposed modification to batch perceptron.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_15"><head>Fig. 4 .</head><label>4</label><figDesc>Fig. 4. Distribution of the sizes of categories in the train dataset</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_16"><head>2</head><label>2</label><figDesc></figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_17"><head>1</head><label></label><figDesc>Dept. of Computer Science LUMS School of Science and Engineering Lahore, Pakistan 2 Dept. of Computer Science</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_18"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Number of tags assigned to posts by users</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_19"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. Discriminative clustering convergence curves (clustering posts based on tags)</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_20"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Tags recommended based on the set of users who have annotated the query resource.</figDesc><graphic coords="101,160.99,183.44,276.67,161.57" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_21"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. Tags recommended based on the set of users who have annotated the query resource and users in the immediate neighborhood of the direct user set.</figDesc><graphic coords="101,160.99,470.17,276.67,141.42" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_22"><head>Fig. 3 .Fig. 4 .</head><label>34</label><figDesc>Fig.3. F-Measure values for Direct user approach and Extended user approach (λ=0.9)</figDesc><graphic coords="103,126.40,373.15,345.84,243.20" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_23"><head>Fig. 5 .</head><label>5</label><figDesc>Fig. 5. F-Measure values for Most Popular Tags per Resource approach and Most Popular and Recent Tags per Resource approach (λ=0.9)</figDesc><graphic coords="105,126.40,140.78,345.82,252.14" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_24"><head>Fig. 6 .</head><label>6</label><figDesc>Fig. 6. F-Measure values for Most Popular Tags per Resource approach and Most Popular and Recent Tags per Resource approach (λ=0.95)</figDesc><graphic coords="106,126.40,140.78,345.82,251.83" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_25"><head>D</head><label></label><figDesc>: set of all documents (resources) such as bookmarks or BibTex references. EC(k, d): extraction count of keyword k in document d. MC(k, d): matching count of keyword k with one of the tags of document d. 5 TEC(d): extraction count of all the keywords in document d.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_26"><head>D:</head><label></label><figDesc>set of all documents (resources) satisfying the same document condition with the present document d. TC(k, d): 1 if document d has keyword k; 0 otherwise. Accuracy Weight from Resource Set, AW RS (k) = ∑ TC(, ).∈</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_27"><head></head><label></label><figDesc>=  ∪  ∪  ( NW DS (ek) = AW DS (ek) / NW RS (ek) = AW RS (ek) / NW US (ek) = AW US (ek) / We also added tag frequency information, denoting how many times a tag was annotated during the training period. TFR(ek) = ∑ TagCount ∈ where TagCount(t, d) denotes the number of document d. T and D (resources), respectively.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_28"><head>Figure 1 . 5 </head><label>15</label><figDesc>Figure 1. Performance information sources.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_29"><head>Figure 2 .</head><label>2</label><figDesc>Figure 2. Performance comparison among different weighting schemes.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_30"><head>Figure 3 .Figure 4 . 8 .</head><label>348</label><figDesc>Figure 3. Effect of candidate elimination on</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_31"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. The parallel architecture of ARKTiS.</figDesc><graphic coords="123,178.28,140.78,242.07,121.72" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_32"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. Sequential processing of textual contents</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_33"><head>Table 4 .</head><label>4</label><figDesc>Top 5 stop words in tags of Cleaned Dump &amp; Post Core dataset top 5 stop words and their frequency in tags Cleaned Dump all:3105 of:1414 and:1227 best:1124 three:1081 c:806 Post Core all:655 open:211 c:165 best:152 work:77</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_34"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Performance of Selected Models</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_35"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. Performance of Selected Models</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_36"><head>Fig. 1 .Fig. 2 .</head><label>12</label><figDesc>Fig.1. Cummulative frequency distribution of resources and users for BibTeX (left) and bookmark (right) data. Much steeper curve for resources shows that we are much less likely to find a rich resource profile, comparing to the profiles of users.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_37"><head>Fig. 3 .</head><label>3</label><figDesc>Fig. 3. Data flow in proposed tag recommendation system.</figDesc><graphic coords="166,126.40,140.78,345.83,208.61" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_38"><head>Fig. 4 .</head><label>4</label><figDesc>Fig. 4. Precision and recall of proposed tag recommendation system and intermediate steps. Test data was divided into BibTeX and bookmark part.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_39"><head>2. 7</head><label>7</label><figDesc>Group 2: Variations for Task 1</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_40"><head>Algorithm 2 3 :</head><label>23</label><figDesc>The single-concept based keyword extraction 1: Let column c of W (Wc) be concept c and d be a document 2: Let rel(d, c) = d T Wc d Wc Find the most relevant concept to the document d, i.e., concept c where c = argmax c (rel(d, c)) 4: Scale terms existed in the document based on the most relevant concept, i.e., d1i = w ic di d2j = t jc dj 5: Combine the normalized terms: d = (1 − α) (d1/ d1 ) + α(d2/ d2 ) 6: Select first n non zero terms of the ranked d as keywords</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_41"><head>Algorithm 3 3 : 4 :</head><label>334</label><figDesc>The multi-concept based keyword extraction 1: Let column c of W (Wc) be concept c and d be a document 2: Let rel(d, c) = d T Wc d Wc Scale terms existed in the document using the concepts, i.e., Combine the normalized terms: d = (1 − α) (d1/ d1 ) + α(d2/ d2 ) 5: Select first n non zero terms of the ranked d as keywords</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_43"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Performance comparison of single-and multi-concept approach</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_45"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. Performance comparison of one-and two-level learning hierarchy approach</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_46"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Architecture of STaR</figDesc><graphic coords="220,143.41,140.78,311.82,175.56" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_47"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. Retrieving of Similar Resources</figDesc><graphic coords="222,171.76,140.78,255.12,212.95" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_48"><head>Fig. 3 .</head><label>3</label><figDesc>Fig. 3. Description of the process performed by the Tag Extractor</figDesc><graphic coords="224,126.40,140.78,240.94,128.79" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_49"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. The procedure of Feature-Driven Tagging.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_50"><head></head><label></label><figDesc>Fig. 2. F-Measure performance for different methods in all collections. For every method, at most 5 tags are recommended. The three methods to compare are only using TF-IDF, linearly combining results of TF-IDF &amp; association rules, and common &amp; combine</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_51"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. Precision/Recall curves for various recommenders on the provided test data.The curve of the U CL recommender appears "shorter" as this recommender suggests a variable number of tags.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_52"><head>Fig. 3 .</head><label>3</label><figDesc>Fig. 3. Relative number of bookmarks per user in test and training data. The correlation between the user participation in the test set and the trained distribution is rather low (ρ = 0.24).</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_54"><head></head><label></label><figDesc>Description T the collection of tags posted in the training data R the collection of resources posted in the training data (grouped by the url_hash or simhash1 ) U the collection of users who posted tags in the training data D training data set containing tagged resources. D={(w i , t i , u i ,)}, which represents a set of pairs of resources and users, with the assigned tags by the corresponding users. D' The test data set containing resources and users. D'={(r j , u i )} {i,j} . Note that: 1) either the user u i or the resource r j may not appear in the training data set. N d number of word tokens in the d ∈ D T d number of tags posted by user u to resource r in d ∈ D 𝐰 d vector form of word tokens in d ∈ D 𝐭 d vector form of tags in d ∈ D u d the user in d ∈ D T (u,r)the set of tags that will be recommended for a given user u and a given resource r z hidden topic layer in ACT model</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_55"><head></head><label></label><figDesc>1) model one only calculates the scores of tags in the candidate set but model two 𝐰 𝟏 ← words in the candidate set C max_score1 ← max w ∈w 1 score1[w] score2 t = P t P(u|t) P(w|t) w∈r T u, r ≔ argmax t∈T n score[t] Input: a given resource r and a given user u and the result of ATC model P(t),P(w|t) and P(u|t) for all tags, users, words in the training file. Output: T (u, r)the set of recommended tags begin //Model one foreach w ∈ 𝐰 1 do 𝑠𝑐𝑜𝑟𝑒1 𝑤 = P 1 w r + P 2 w r end //Model two foreach t ∈ T do end max_score2 ← max t∈T score2[t] //Combination foreach t ∈ 𝐰 𝟏 ∪ T do score[t] = score1[t]+score2[t]*max_score1/max_score2 end end calculates all the tags t ∈ T.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_56"><head>Fig. 1</head><label>1</label><figDesc>Fig.1 Recall and precision of tag recommendation system</figDesc><graphic coords="294,125.98,150.73,345.75,235.50" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_57"><head></head><label></label><figDesc>tags posted in the training data R the collection of resources posted in the training data (grouped by the url_hash or simhash1 ) U the collection of users who posted tags in the training data D training data set containing tagged resources. D={(r j , t i, u i )}, which represents a set of pairs of resources and users, with the assigned tags by the corresponding users. D' The test data set containing resources and users. D'={(r j , u i )}. Note that: the user u i , the resource r j and the original tags posted by u i to r i appear in the training dataset. N d number of word tokens in the d ∈ D T r number of tags posted to resource r 𝐭 d vector form of tags in d ∈ D u d the user in d ∈ D C(u, r)</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_58"><head></head><label></label><figDesc>resources. E = u. t , t, r , u, r {r, t, u} ∈ D} and each edge {u, t} ∈ E has a weight | r ∈ R r, t, u ∈ D |, each edge t, r ∈ E has a weight | u ∈ U|{r, t, u} ∈ D | and each edge u, r ∈ E has a weight | t ∈ T|{r, t, u} ∈ D | . After having the graph format of the posts, we can spread the weight like PageRank as follows:</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_59"><head>Algorithm 1 :</head><label>1</label><figDesc>The FolkRank algorithm used in our tag recommendation system Input: the graph information of the training file, i.e. G = (V, E) where V =T ∪ R ∪ U and E = u. t , t, r , u, r {r, t, u} ∈ D}, the adjacency matrix A, the given resource r and the given user u. Output: the ranking w of all tags ∈ T begin //Initialize foreach t ∈ T, r ∈ R and u ∈ U do w 0 [t] = w 1 [t]=1,w 0 [r]= w 1 [r]=2 and w 0 [u] = w 1 [u] =2 end foreach t ∈ T, r ∈ R and u ∈ U do p[t]=p[r]=p[u]=1 end p[r] = 1+|R| p[u]= 1+|U| d = 0.625 //iteration for 𝑤 1 repeat w 1 = dAw 1 + (1 − d)p until convergence //iteration for 𝑤 0 repeat w 0 = Aw 0 until convergence w = w 1 −</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0"><head></head><label></label><figDesc></figDesc><graphic coords="104,126.40,140.78,345.84,245.72" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0"><head></head><label></label><figDesc></figDesc><graphic coords="304,133.48,431.98,307.50,208.50" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 .</head><label>1</label><figDesc>Characteristics of 2-core BibSonomy.</figDesc><table><row><cell>dataset</cell><cell>|U |</cell><cell>|R|</cell><cell>|T |</cell><cell>|Y |</cell><cell>|X|</cell></row><row><cell cols="6">BibSonomy 1,185 22,389 13,276 253,615 64,406</cell></row><row><cell></cell><cell>dataset</cell><cell cols="3">|U | |R| |Xtest|</cell><cell></cell></row><row><cell></cell><cell>Holdout</cell><cell cols="3">292 788 800</cell><cell></cell></row><row><cell></cell><cell cols="4">Challenge test 136 667 778</cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 .</head><label>2</label><figDesc>Characteristics of the holdout set and the challenge test dataset.</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 1 .</head><label>1</label><figDesc>Meta-information available in BibSonomy system about two different bookmarks: a web page and a scientific publication.</figDesc><table><row><cell>URL</cell><cell>http://www.adammathes.com/academic/computer-mediated-communication/ folksonomies.html</cell></row><row><cell cols="2">Description Folksonomies -Cooperative Classification and Communication Through Shared</cell></row><row><cell></cell><cell>Metadata</cell></row><row><cell>Extended</cell><cell>General overview of tagging and folksonomies. Difference between controlled</cell></row><row><cell></cell><cell>vocabularies, author and user tagging. Advantages and shortcomings of</cell></row><row><cell></cell><cell>folksonomies</cell></row><row><cell>Title</cell><cell>Semantic Modelling of User Interests Based on Cross-Folksonomy Analysis</cell></row><row><cell>Author</cell><cell>M. Szomszor and H. Alani and I. Cantador and K. O'hara and N. Shadbolt</cell></row><row><cell>Booktitle</cell><cell>Proceedings of the 7th International Semantic Web Conference (ISWC 2008)</cell></row><row><cell>Journal</cell><cell>The Semantic Web -ISWC 2008</cell></row><row><cell>Pages</cell><cell>632-648</cell></row><row><cell>URL</cell><cell>http://dx.doi.org/10.1007/978-3-540-88564-1_40</cell></row><row><cell>Year</cell><cell>2008</cell></row><row><cell>Month</cell><cell>October</cell></row><row><cell>Location</cell><cell>Karlsruhe, Germany</cell></row><row><cell>Abstract</cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 3 .</head><label>3</label><figDesc>Extracted keywords, generated query, and retrieved similar bookmarks for the example input bookmark.</figDesc><table><row><cell>Input bookmark: A</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Multilayer Ontology-based Hybrid Recommendation Model</head><label></label><figDesc></figDesc><table><row><cell>Keywords</cell><cell>multilayer, ontology, hybrid, recommendation, configwork, aicom,</cell></row><row><cell></cell><cell>ai, communication, user, preference, semantic, concept, domain,</cell></row><row><cell></cell><cell>ontology, item, space, way, cluster, similarity, individual, layer,</cell></row><row><cell></cell><cell>community, interest</cell></row><row><cell>Query</cell><cell>recommendation^0.125, ontology^0.09375, concept^0.0625,</cell></row><row><cell></cell><cell>hybrid^0.0625, item^0.0625, layer^0.0625, multilayer^0.0625,</cell></row><row><cell></cell><cell>semantic^0.0625, user^0.0625, aicom^0.03125, cluster^0.03125,</cell></row><row><cell></cell><cell>configwork^0.03125, individual^0.03125, interest^0.03125,</cell></row><row><cell></cell><cell>communication^0.03125, community^0.03125, preference^0.03125,</cell></row><row><cell></cell><cell>similarity^0.03125, space^0.03125, way^0.03125</cell></row><row><cell>Similar bookmarks</cell><cell>• Improving Recommendation Lists Through Topic</cell></row><row><cell></cell><cell>Diversification</cell></row><row><cell></cell><cell>• Item-Based Collaborative Filtering Recommendation</cell></row><row><cell></cell><cell>Algorithms</cell></row><row><cell></cell><cell>• Probabilistic Models for Unified Collaborative and Content-</cell></row><row><cell></cell><cell>Based Recommendation in Sparse-Data Environments</cell></row><row><cell></cell><cell>• Automatic Tag Recommendation for the Web 2.0 Blogosphere</cell></row><row><cell></cell><cell>using Collaborative Tagging and Hybrid ANN semantic</cell></row><row><cell></cell><cell>structures</cell></row><row><cell></cell><cell>• PIMO -a Framework for Representing Personal Information</cell></row><row><cell></cell><cell>Models</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 4 .</head><label>4</label><figDesc>Weighted subset of tags retrieved from the list of bookmarks that are similar to the example input bookmark.</figDesc><table><row><cell>Input bookmark:</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_6"><head>A Multilayer Ontology-based Hybrid Recommendation Model</head><label></label><figDesc></figDesc><table><row><cell>Related tag</cell><cell>Weight</cell><cell>Related tag</cell><cell>Weight</cell><cell>Related tag</cell><cell>Weight</cell></row><row><cell>recommender</cell><cell cols="2">10.538 clustering</cell><cell cols="2">2.013 dataset</cell><cell>0.871</cell></row><row><cell>recommendation</cell><cell cols="4">6.562 recommendersystems 1.669 evaluation</cell><cell>0.786</cell></row><row><cell>collaborative</cell><cell>5.142 web</cell><cell></cell><cell cols="2">1.669 suggestion</cell><cell>0.786</cell></row><row><cell>filtering</cell><cell cols="2">5.142 information</cell><cell cols="2">1.539 semantics</cell><cell>0.786</cell></row><row><cell>collaborativefiltering</cell><cell>3.585 ir</cell><cell></cell><cell>1.378 tag</cell><cell></cell><cell>0.786</cell></row><row><cell>ecommerce</cell><cell cols="2">3.138 retrieval</cell><cell cols="2">1.378 tagging</cell><cell>0.786</cell></row><row><cell>personalization</cell><cell cols="5">3.138 contentbasedfiltering 1.006 knowledgemanagement 0.290</cell></row><row><cell>cf</cell><cell cols="2">2.757 ontologies</cell><cell cols="2">1.006 network</cell><cell>0.290</cell></row><row><cell>semantic</cell><cell cols="2">2.745 ontology</cell><cell>1.006 neural</cell><cell></cell><cell>0.290</cell></row><row><cell>semanticweb</cell><cell cols="2">2.259 userprofileservices</cell><cell cols="2">1.006 neuralnetwork</cell><cell>0.290</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_7"><head>Table 5 .</head><label>5</label><figDesc>Final tag recommendations for the example input bookmark.</figDesc><table><row><cell>Input bookmark: A</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_8"><head>Multilayer Ontology-based Hybrid Recommendation Model</head><label></label><figDesc></figDesc><table><row><cell>Tag 1 recommender</cell></row><row><cell>Tag 2 collaborative</cell></row><row><cell>Tag 3 filtering</cell></row><row><cell>Tag 4 semanticweb</cell></row><row><cell>Tag 5 personalization</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_9"><head></head><label></label><figDesc>Forming part of the ECML PKDD 2009 Discovery Challenge, two experimental tasks have been designed to evaluate the tag recommendations. Both of them get the same dataset for training, a snapshot of BibSonomy system until December 31st 2008, but different test datasets: • Task 1. The test data contains bookmarks, whose user, resource or tags are not contained in the training data. • Task 2. The test data contains bookmarks, whose user, resource or tags are all contained in the training data.</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_10"><head>Table 6 .</head><label>6</label><figDesc>ECML PKDD 2009 Discovery Challenge dataset.</figDesc><table><row><cell></cell><cell></cell><cell>Web pages</cell><cell>Scientific publications</cell><cell>All bookmarks</cell></row><row><cell></cell><cell>users</cell><cell>2679</cell><cell>1790</cell><cell></cell></row><row><cell></cell><cell>resources</cell><cell>263004</cell><cell>158924</cell><cell>421928</cell></row><row><cell>Training</cell><cell>tags</cell><cell>56424</cell><cell>50855</cell><cell>93756</cell></row><row><cell></cell><cell>tas</cell><cell>916469</cell><cell>484635</cell><cell>1401104</cell></row><row><cell></cell><cell>tas/resource</cell><cell>3.48</cell><cell>3.05</cell><cell>3.32</cell></row><row><cell></cell><cell>users</cell><cell>891</cell><cell>1045</cell><cell></cell></row><row><cell></cell><cell>resources</cell><cell>16898</cell><cell>26104</cell><cell>43002</cell></row><row><cell>Test (task 1)</cell><cell>tags</cell><cell>14395</cell><cell>24393</cell><cell>34051</cell></row><row><cell></cell><cell>tas</cell><cell>64460</cell><cell>99603</cell><cell>164063</cell></row><row><cell></cell><cell>tas/resource</cell><cell>3.81</cell><cell>3.82</cell><cell>3.82</cell></row><row><cell></cell><cell>users</cell><cell>91</cell><cell>81</cell><cell>136</cell></row><row><cell></cell><cell>resources</cell><cell>431</cell><cell>347</cell><cell>778</cell></row><row><cell>Test (task 2)</cell><cell>tags</cell><cell>587</cell><cell>397</cell><cell>862</cell></row><row><cell></cell><cell>tas</cell><cell>1465</cell><cell>1139</cell><cell></cell></row><row><cell></cell><cell>tas/resource</cell><cell>3.40</cell><cell>3.28</cell><cell>4.35</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_11"><head>Table 7 .</head><label>7</label><figDesc>Average recall, precision and F-measure values obtained in tasks 1 and 2 of ECML PKDD 2009 Discovery Challenge for different numbers of recommended tags.</figDesc><table><row><cell></cell><cell>umber of recommended tags</cell><cell>Recall</cell><cell>Precision</cell><cell>F-measure</cell></row><row><cell></cell><cell>1</cell><cell>0.0593</cell><cell>0.1810</cell><cell>0.0894</cell></row><row><cell></cell><cell>2</cell><cell>0.0910</cell><cell>0.1453</cell><cell>0.1120</cell></row><row><cell>Task 1</cell><cell>3</cell><cell>0.1131</cell><cell>0.1233</cell><cell>0.1179</cell></row><row><cell></cell><cell>4</cell><cell>0.1309</cell><cell>0.1091</cell><cell>0.1190</cell></row><row><cell></cell><cell>5</cell><cell>0.1454</cell><cell>0.0991</cell><cell>0.1179</cell></row><row><cell></cell><cell>1</cell><cell>0.1454</cell><cell>0.4190</cell><cell>0.2159</cell></row><row><cell></cell><cell>2</cell><cell>0.2351</cell><cell>0.3477</cell><cell>0.2805</cell></row><row><cell>Task 2</cell><cell>3</cell><cell>0.2991</cell><cell>0.3059</cell><cell>0.3025</cell></row><row><cell></cell><cell>4</cell><cell>0.3462</cell><cell>0.2716</cell><cell>0.3044</cell></row><row><cell></cell><cell>5</cell><cell>0.3916</cell><cell>0.2518</cell><cell>0.3065</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_12"><head>Table 1 .</head><label>1</label><figDesc>Algorithm of rank the candidate tags Input: candidate tags {CT1, CT2, ..., CTn} Output: top-k tags {CT ′ 1 , CT ′ 2 , ..., CT ′</figDesc><table /><note>k } 1. Extract feature x = {xi}(i = 1, 2, ..., n) for a sequence of candidate tags CT {Pij } = {CT1, CT2, ..., CTn}. 2. Rank the features using the learned ranking model as {CT ′ 1 , CT ′ 2 , ..., CT ′ n }. 3. select top-k tags {CT ′ 1 , CT ′ 2 , ..., CT ′ k } as recommending tags.</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_13"><head>Table 2 .</head><label>2</label><figDesc>Statistics of posts on recovered 08's dataset</figDesc><table><row><cell>Post in recovered training data</cell><cell>234,134</cell><cell>BOOKMARK 184,655 BIBTEX 49,479</cell></row><row><cell>Post in recovered test data</cell><cell>63,192</cell><cell>BOOKMARK 20,647 BIBTEX 42,545</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_14"><head>Table 3 .</head><label>3</label><figDesc>Statistics of posts according to their user and resource statusData format description The dataset used in experiments is released by ECML. The data consists of three tables: TAS table, BOOKMARK table and BIBTEX table.Table4is a description of the fields of the three tables. Only the fields we used in experiments are listed in the table.</figDesc><table><row><cell>Users in recovered test data appear in recovered training data</cell><cell>265</cell></row><row><cell>Users in recovered test data do not appear in recovered training data</cell><cell>225</cell></row><row><cell>Resources in recovered test data appear in recovered training data</cell><cell>1230</cell></row><row><cell cols="2">Resources in recovered test data do not appear in recovered training data 61970</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_15"><head>Table 4 .</head><label>4</label><figDesc>Data fields of TAS, BOOKMARK and BIBTEX</figDesc><table><row><cell cols="2">Table name Fields name</cell></row><row><cell>TAS</cell><cell>user, tag, content id, content type, date</cell></row><row><cell cols="2">BOOKMARK content id (matches tas.content id) ,url</cell></row><row><cell></cell><cell>description ,extended ,description ,date ,bibtex</cell></row><row><cell>BIBTEX</cell><cell>content id (matches tas.content id) ,simhash1 (hash for duplicate detection</cell></row><row><cell></cell><cell>among users) ,title</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_16"><head>Table 5 .</head><label>5</label><figDesc>Example results of data preprocess</figDesc><table><row><cell>Before data preprocess</cell><cell>After data preprocess</cell></row><row><cell>Ben Mezrich: the telling of a true</cell><cell>ben mezrich telling true story</cell></row><row><cell>story</cell><cell></cell></row><row><cell>{XQ}uery 1.0: An {XML} Query</cell><cell>xquery 1.0 xml query language w3c</cell></row><row><cell>Language, {W3C} Working Draft</cell><cell>working draft</cell></row><row><cell cols="2">some resources of posts exist in the training data (2%) and others do not exist</cell></row><row><cell>in the training data (98%).</cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_17"><head>Table 6 .</head><label>6</label><figDesc>Simplified symbols EUER post Existed user existed resource post EUNR post Existed user non-existed resource post NUER post Non-existed user existed resource post NUNR post Non-existed user non-existed resource post</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_18"><head>Table 7 .</head><label>7</label><figDesc>Distribution of different categories of BOOKMARK posts in test dataset</figDesc><table><row><cell>Category</cell><cell cols="2">Posts number ratio</cell></row><row><cell cols="2">EUER post 621</cell><cell>3.01%</cell></row><row><cell cols="2">EUNR post 17099</cell><cell>82.80%</cell></row><row><cell cols="2">NUER post 346</cell><cell>1.68%</cell></row><row><cell cols="2">NUNR post 2585</cell><cell>12.52%</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_19"><head>Table 8 .</head><label>8</label><figDesc>Distribution of different categories of BIBTEX posts in test dataset</figDesc><table><row><cell>Category</cell><cell cols="2">Posts number ratio</cell></row><row><cell cols="2">EUER post 164</cell><cell>0.39%</cell></row><row><cell cols="2">EUNR post 2532</cell><cell>5.95%</cell></row><row><cell cols="2">NUER post 99</cell><cell>0.23%</cell></row><row><cell cols="2">NUNR post 39754</cell><cell>93.43%</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_20"><head>Table 9 .</head><label>9</label><figDesc>Statistics of the tags from 3 sources of BOOKMARK Post</figDesc><table><row><cell>Total tags</cell><cell>56267</cell></row><row><cell cols="2">Tags from terms of description 5253</cell></row><row><cell>Tags from terms of URL</cell><cell>1353</cell></row><row><cell cols="2">Tags from user's previous tags 29672</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_21"><head>Table 10 .</head><label>10</label><figDesc>Statistics of the tags from 3 sources of BIBTEX Post</figDesc><table><row><cell>Total tags</cell><cell>95782</cell></row><row><cell>Tags from terms of title</cell><cell>41801</cell></row><row><cell>Tags from terms of URL</cell><cell>547</cell></row><row><cell cols="2">Tags from user's previous tags 5377</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_22"><head>Table 11 .</head><label>11</label><figDesc>Some of the features for ranking SVM model for BOOKMARK Feature1 Candidate tag's TF (term frequency) in post's description terms. Feature2 Candidate tag's TF in post's URL terms. Feature3 Candidate tag's TF in post's extended description terms. Feature4 Candidate tag's TF in T {Rj } (tags assigned to the post of the same URL in the training data). Feature5 Candidate tag's TF in T {Ui} (tags assigned previously by user in the training data.) Feature6 Times of candidate tag being assigned as a tag in the training data.</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_23"><head>Table 12 .</head><label>12</label><figDesc>Individual and overall Performance on BOOKMARK posts</figDesc><table><row><cell>Post category</cell><cell>Recall Precision F1-value ratio</cell></row><row><cell>EUER Post</cell><cell>0.369699 0.394973 0.381918 3.01%</cell></row><row><cell>EUNR Post</cell><cell>0.046591 0.053739 0.04991 82.80%</cell></row><row><cell>NUER Post</cell><cell>0.160883 0.255652 0.197487 1.68%</cell></row><row><cell>NUNR Post</cell><cell>0.069158 0.106366 0.083819 12.52%</cell></row><row><cell cols="2">overall-performance on BOOKMARK 0.061067 0.073997 0.066633</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_24"><head>Table 13 .</head><label>13</label><figDesc>Individual and overall Performance on BIBTEX posts</figDesc><table><row><cell>Post category</cell><cell>Recall</cell><cell>Precision F1-value ratio</cell></row><row><cell>EUER Post</cell><cell cols="2">0.4219356 0.3472393 0.3809605 0.39%</cell></row><row><cell>EUNR Post</cell><cell cols="2">0.2250226 0.1628605 0.1889605 5.95%</cell></row><row><cell>NUER Post</cell><cell cols="2">0.5667162 0.3715986 0.4488706 0.23%</cell></row><row><cell>NUNR Post</cell><cell cols="2">0.3561221 0.1603686 0.2211494 93.43%</cell></row><row><cell cols="3">overall-performance on BIBTEX 0.349063 0.161732 0.220381</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_25"><head>Table 14 .</head><label>14</label><figDesc>Overall performance on test dataset using ranking SVM model</figDesc><table><row><cell cols="2">Recall Precision F1-value</cell></row><row><cell>0.153 0.185</cell><cell>0.167</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_26"><head>Table 15 .</head><label>15</label><figDesc>Overall performance on test dataset adding content similarity based KNN model</figDesc><table><row><cell>Recall Precision F1-value</cell></row><row><cell>0.323828 0.200926 0.238803</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_27"><head>Table 16 .</head><label>16</label><figDesc>Different categories of BOOKMARK posts in 09s test dataset for Task 1</figDesc><table><row><cell>Category</cell><cell cols="2">Posts number ratio</cell></row><row><cell cols="2">EUER Post 821</cell><cell>4.86%</cell></row><row><cell cols="2">EUNR Post 10622</cell><cell>62.86%</cell></row><row><cell cols="2">NUER Post 872</cell><cell>5.16%</cell></row><row><cell cols="2">NUNR Post 4583</cell><cell>27.12%</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_28"><head>Table 17 .</head><label>17</label><figDesc>Different categories of BIBTEX posts in 09s test dataset for Task 1</figDesc><table><row><cell>Category</cell><cell cols="2">Posts number ratio</cell></row><row><cell cols="2">EUER Post 365</cell><cell>1.40%</cell></row><row><cell cols="2">EUNR Post 9287</cell><cell>35.71%</cell></row><row><cell cols="2">NUER Post 591</cell><cell>2.27%</cell></row><row><cell cols="2">NUNR Post 15761</cell><cell>60.61%</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_29"><head>Table 18 .</head><label>18</label><figDesc>Performance on 09's dataset @5</figDesc><table><row><cell cols="3">Task No. Submission ID Precision Recall F1-value</cell></row><row><cell>1</cell><cell>67797</cell><cell>0.162478 0.146582 0.154121</cell></row><row><cell>2</cell><cell>13651</cell><cell>0.31622 0.222065 0.260908</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_30"><head>Table 1 .</head><label>1</label><figDesc>The most common tags in Delicious and Bibsonomy</figDesc><table><row><cell></cell><cell>Tag</cell><cell>Delicious 1</cell><cell>BibSonomy 2</cell></row><row><cell>1</cell><cell>design</cell><cell>1.69%</cell><cell>27</cell></row><row><cell>2</cell><cell>blog</cell><cell>1.29%</cell><cell>13</cell></row><row><cell>3</cell><cell>tools</cell><cell>1.05%</cell><cell>10</cell></row><row><cell>4</cell><cell>software</cell><cell>0.96%</cell><cell>4</cell></row><row><cell>5</cell><cell>webdesign</cell><cell>0.92%</cell><cell>54</cell></row><row><cell>6</cell><cell>programming</cell><cell>0.89%</cell><cell>5</cell></row><row><cell>7</cell><cell>tutorial</cell><cell>0.85%</cell><cell>44</cell></row><row><cell>8</cell><cell>art</cell><cell>0.75%</cell><cell>83</cell></row><row><cell>9</cell><cell>reference</cell><cell>0.72%</cell><cell>33</cell></row><row><cell>10</cell><cell>video</cell><cell>0.72%</cell><cell>3</cell></row><row><cell>11</cell><cell>inspiration</cell><cell>0.71%</cell><cell>587</cell></row><row><cell>12</cell><cell>music</cell><cell>0.66%</cell><cell>25</cell></row><row><cell>13</cell><cell>web2.0</cell><cell>0.65%</cell><cell>7</cell></row><row><cell>14</cell><cell>education</cell><cell>0.63%</cell><cell>17</cell></row><row><cell>15</cell><cell>photography</cell><cell>0.52%</cell><cell>166</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_31"><head>Table 2 .</head><label>2</label><figDesc>Results of searching first ten feeds for "Linux" keyword in Google Reader.</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_32"><head>Table 3 .</head><label>3</label><figDesc>Co-occurances of pairs of tags, occurances and normalized Jaccard coefficient.</figDesc><table><row><cell>Tag 1</cell><cell>Tag 2</cell><cell>Éˮ # ˮ $ É</cell><cell>|t 1 |</cell><cell>|t 2 |</cell><cell>Éˮ # ˮ $ É Éˮ # ˮ $ É</cell></row><row><cell>ccp</cell><cell>jrr</cell><cell>4294</cell><cell>4294</cell><cell>4294</cell><cell>1.00</cell></row><row><cell>algorithms</cell><cell>genetic</cell><cell>5775</cell><cell>6220</cell><cell>5888</cell><cell>0.91</cell></row><row><cell>aaaemulation-topgames</cell><cell>emulation-videogames</cell><cell>3653</cell><cell>3653</cell><cell>4576</cell><cell>0.80</cell></row><row><cell>emulationgames</cell><cell>emulation-videogames</cell><cell>4576</cell><cell>6055</cell><cell>4576</cell><cell>0.76</cell></row><row><cell>aaaemulation-topgames</cell><cell>classicemulated-remakeretrogames</cell><cell>2472</cell><cell>3653</cell><cell>2472</cell><cell>0.68</cell></row><row><cell>aaaemulation-topgames</cell><cell>emulationgames</cell><cell>3653</cell><cell>3653</cell><cell>6055</cell><cell>0.60</cell></row><row><cell>classicemulated-remakeretrogames</cell><cell>emulation-videogames</cell><cell>2472</cell><cell>2472</cell><cell>4576</cell><cell>0.54</cell></row><row><cell>genetic</cell><cell>programming</cell><cell>5262</cell><cell>5888</cell><cell>9491</cell><cell>0.52</cell></row><row><cell>journal</cell><cell>medical</cell><cell>1693</cell><cell>2566</cell><cell>2448</cell><cell>0.51</cell></row><row><cell>algorithms</cell><cell>programming</cell><cell>5303</cell><cell>6220</cell><cell>9491</cell><cell>0.51</cell></row><row><cell>classicemulated-remakeretrogames</cell><cell>emulationgames</cell><cell>2472</cell><cell>2472</cell><cell>6055</cell><cell>0.41</cell></row><row><cell>book</cell><cell>nlp</cell><cell>1230</cell><cell>2614</cell><cell>2027</cell><cell>0.36</cell></row><row><cell>education</cell><cell>learning</cell><cell>2143</cell><cell>5021</cell><cell>4751</cell><cell>0.28</cell></row><row><cell>media</cell><cell>texts</cell><cell>1998</cell><cell>7149</cell><cell>2012</cell><cell>0.28</cell></row><row><cell>analysis</cell><cell>data</cell><cell>1187</cell><cell>3352</cell><cell>2589</cell><cell>0.25</cell></row><row><cell>folksonomy</cell><cell>tagging</cell><cell>1027</cell><cell>2561</cell><cell>3083</cell><cell>0.22</cell></row><row><cell>emulationgames</cell><cell>zzztosort</cell><cell>2844</cell><cell>6055</cell><cell>11839</cell><cell>0.19</cell></row><row><cell>audio</cell><cell>music</cell><cell>919</cell><cell>1857</cell><cell>4142</cell><cell>0.18</cell></row><row><cell>howto</cell><cell>tutorial</cell><cell>850</cell><cell>2876</cell><cell>2798</cell><cell>0.18</cell></row><row><cell>bookmarks</cell><cell>indexforum</cell><cell>9164</cell><cell>52795</cell><cell>9183</cell><cell>0.17</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_35"><head>Table 1 .</head><label>1</label><figDesc>Bibsonomy datasets.</figDesc><table><row><cell></cell><cell cols="2">Complete PostCore(2)</cell></row><row><cell>Users</cell><cell>3,617</cell><cell>253,615</cell></row><row><cell>URLs</cell><cell>235,328</cell><cell>41,268</cell></row><row><cell>BibTeXs</cell><cell>143,050</cell><cell>22,852</cell></row><row><cell>Tags</cell><cell>93,756</cell><cell>1,185</cell></row><row><cell cols="2">Tag Assignments 1,401,104</cell><cell>14,443</cell></row><row><cell>Bookmark Posts</cell><cell>263,004</cell><cell>7,946</cell></row><row><cell>BibTeX Posts</cell><cell>158,924</cell><cell>13,276</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_36"><head>Table 1 .</head><label>1</label><figDesc>Statistics for categories and documents in datasets</figDesc><table><row><cell></cell><cell cols="2">Train dataset Test dataset</cell></row><row><cell>Number of Documents</cell><cell>64,120</cell><cell>778</cell></row><row><cell>Number of Categories</cell><cell>13,276</cell><cell>-</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_37"><head>Table 1 .</head><label>1</label><figDesc>Post-core at level 2 data statistics</figDesc><table><row><cell></cell><cell cols="3">Avg Min Max Std. Deviation</cell></row><row><cell cols="2">No. of tags per post 4</cell><cell>1 81</cell><cell>3.3</cell></row><row><cell cols="3">No. of posts per user 54 2 2031</cell><cell>162.9</cell></row><row><cell cols="3">No. of tags per user 62 1 4711</cell><cell>214.5</cell></row><row><cell>Frequency of tags</cell><cell cols="2">19 2 4474</cell><cell>106.9</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_38"><head>Table 2 .</head><label>2</label><figDesc>Performance of discriminative clustering of posts using the tags assigned to them (postcore at level 2 data)</figDesc><table><row><cell>K</cell><cell>10 50 100 200 300</cell></row><row><cell>Act. Clusters</cell><cell>10 48 95 189 274</cell></row><row><cell cols="2">Av. Precision (%) 12.5 19.2 22.3 25.2 26.9</cell></row><row><cell>Av. Recall (%)</cell><cell>21.0 32.8 38.6 45.9 48.7</cell></row><row><cell cols="2">Av. F1-score (%) 13.7 21.4 25.0 28.7 30.6</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_39"><head>Table 3 .</head><label>3</label><figDesc>Top tags for selected clusters (K = 200)</figDesc><table><row><cell cols="2">/ No. Top Discriminating Tags</cell></row><row><cell>1</cell><cell>svm, ki2007webmining, mining, kernels, textmining, dm, textclassification</cell></row><row><cell>2</cell><cell>windows, freeware, utility, download, utilities, win, shareware</cell></row><row><cell>3</cell><cell>fun, flash, games, game, microfiction, flashfiction, sudden</cell></row><row><cell>4</cell><cell>tag, cloud, tagcloud, tags, folksonomia, tagging, vortragmnchen2008</cell></row><row><cell>5</cell><cell>library, books, archive, bibliothek, catalog, digital, opac</cell></row><row><cell>6</cell><cell>voip, mobile, skype, phone, im, messaging, hones</cell></row><row><cell>7</cell><cell>rss, feeds, aggregator, feed, atom, syndication, opml</cell></row><row><cell>8</cell><cell>bookmarks, bookmark, tags, bookmarking, delicious, diigo, socialbookmarking</cell></row></table><note>each group of posts. Noisy tags are not ranked high in the lists. It is even able to discriminate and group posts of different languages (not shown in this table), especially when clustering is based on content terms. Two valuable characteristics of the discriminative clustering method are its stability and efficiency. The method converges smoothly (Figure2) usually within 15 iteration. More importantly, especially considering the large post by vocabulary sizes involved, is the efficiency of the method. Each iteration of the method completes within 3 minutes, even for the large 107, 122 × 317, 283 data for the content-based clustering of the post-core plus task 1 test data.</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_40"><head>Table 4 .</head><label>4</label><figDesc>Tag recommendation performance (average F1-score percentages) using TG or TM only for original data</figDesc><table><row><cell>K</cell><cell>10 50 100 200 300</cell></row><row><cell cols="2">TG Only (Best Cluster) 6.6 7.4 8.7 8.7 7.2</cell></row><row><cell cols="2">TG Only (Top 3 Clusters) 7.3 8.2 9.5 10.6 9.1</cell></row><row><cell>TM Only (Best Cluster)</cell><cell>6.3</cell></row><row><cell>TM Only (Top 3 Clusters)</cell><cell>7.8</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_41"><head>Table 5 .</head><label>5</label><figDesc>Tag recommendation performance (average F1-score percentages) for processed data (K = 200; prediction based on top 3 clusters). The bottom line shows performance on task 1 test data</figDesc><table><row><cell>Data / Lists)</cell><cell cols="3">TF TG TM TG, TM TG, TM, TU</cell></row><row><cell>Original Contents</cell><cell cols="2">7.0 10.6 7.8 11.5</cell><cell>12.8</cell></row><row><cell>Crawled Contents</cell><cell cols="2">7.0 12.3 10.4 14.3</cell><cell>15.5</cell></row><row><cell cols="3">Crawled+Lemmatized Contents 7.0 11.7 9.7 13.3</cell><cell>14.6</cell></row><row><cell>Task 1 Test Data (Crawled)</cell><cell>1.1 4.9 3.2</cell><cell>5.2</cell><cell>5.4</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_43"><head>Table 1 . Example keywords with high accuracy ratios.</head><label>1</label><figDesc>Some keywords with high accuracy ratio values are shown in Table 1. It should be noted that there exist a large amount of keywords having high AR(k) values and the keywords in Table 1 are a sample from them.</figDesc><table><row><cell cols="3">Limit Condition: AR(k) / (1 + FR(k)) &gt; ∑</cell><cell>∈</cell><cell cols="2">TMC()</cell><cell>⁄</cell><cell>∑</cell><cell>∈</cell><cell>TEC()</cell><cell>.</cell><cell>(3)</cell></row><row><cell>Keywords</cell><cell>Extracted</cell><cell cols="3">Accuracy ratio,</cell><cell cols="5">Frequency ratio,</cell></row><row><cell></cell><cell>columns</cell><cell cols="3">AR(k)</cell><cell></cell><cell></cell><cell cols="3">FR(k)</cell></row><row><cell>nejm</cell><cell>extended</cell><cell></cell><cell cols="2">1.0000</cell><cell></cell><cell></cell><cell></cell><cell cols="2">0.0002579</cell></row><row><cell></cell><cell>description</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>medscape</cell><cell>extended</cell><cell></cell><cell cols="2">1.0000</cell><cell></cell><cell></cell><cell></cell><cell cols="2">0.0001146</cell></row><row><cell></cell><cell>description</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>freebox</cell><cell>description</cell><cell></cell><cell cols="2">1.0000</cell><cell></cell><cell></cell><cell></cell><cell cols="2">0.0000533</cell></row><row><cell>harum</cell><cell>description</cell><cell></cell><cell cols="2">0.9800</cell><cell></cell><cell></cell><cell></cell><cell cols="2">0.0000556</cell></row><row><cell>ldap</cell><cell>url</cell><cell></cell><cell cols="2">0.9354</cell><cell></cell><cell></cell><cell></cell><cell cols="2">0.0000403</cell></row><row><cell>shipyard</cell><cell>description</cell><cell></cell><cell cols="2">0.9146</cell><cell></cell><cell></cell><cell></cell><cell cols="2">0.0002734</cell></row></table><note>TMC(d): sum of MC(k, d) across all the keywords in document d.</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_44"><head>Table 2 . Example keywords on the border in terms of Limit Condition.</head><label>2</label><figDesc>The accuracy weight of each candidate is calculated by multiplying its accuracy ratio and extraction count from the present document as follows.Accuracy Weight from Document Set, AW DS</figDesc><table><row><cell>Keywords</cell><cell>Extracted</cell><cell>Accuracy</cell><cell>Frequency</cell><cell>Difference</cell><cell>Limit</cell></row><row><cell></cell><cell>columns</cell><cell>ratio,</cell><cell>ratio, FR(k)</cell><cell>in Limit</cell><cell>Condition</cell></row><row><cell></cell><cell></cell><cell>AR(k)</cell><cell></cell><cell>Condition</cell><cell>satisfied</cell></row><row><cell>netbib</cell><cell>url</cell><cell cols="2">0.0789 0.0002468</cell><cell>0.0004281</cell><cell>Yes</cell></row><row><cell>guide</cell><cell>url</cell><cell cols="3">0.0778 0.0006510 -0.0007060</cell><cell>No</cell></row><row><cell>media</cell><cell>url</cell><cell cols="3">0.0781 0.0008810 -0.0003974</cell><cell>No</cell></row><row><cell>daily</cell><cell>extended</cell><cell cols="2">0.0602 0.0002744</cell><cell>0.0005867</cell><cell>Yes</cell></row><row><cell></cell><cell>description</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>list</cell><cell>extended</cell><cell cols="2">0.0601 0.0008056</cell><cell>0.0005053</cell><cell>Yes</cell></row><row><cell></cell><cell>description</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>engine</cell><cell>extended</cell><cell cols="3">0.0590 0.0005598 -0.0006156</cell><cell>No</cell></row><row><cell></cell><cell>description</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>tool</cell><cell>description</cell><cell cols="2">0.1279 0.0007647</cell><cell>0.0006509</cell><cell>Yes</cell></row><row><cell>ontologies</cell><cell>description</cell><cell cols="3">0.1271 0.0001312 -0.0000749</cell><cell>No</cell></row><row><cell>corpus</cell><cell>description</cell><cell cols="3">0.1264 0.0000967 -0.0007337</cell><cell>No</cell></row><row><cell cols="6">The average AR(k) values in url, description, and extended description are</cell></row><row><cell cols="6">0.07849973, 0.12715830, and 0.05963773, respectively. We also show some example</cell></row><row><cell cols="6">keywords with low accuracy ratios, which do not satisfy Equation (3) as follows.</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_45"><head>Table 3 . Variation in accuracy from different data columns. represented in bold.</head><label>3</label><figDesc></figDesc><table><row><cell>Keywords</cell></row><row><cell>portal</cell></row><row><cell>tag</cell></row><row><cell>tech</cell></row><row><cell>template</cell></row><row><cell>time</cell></row><row><cell>youtube</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_46"><head>. . Variation in accuracy ratio, AR(k), of the same keywords extracted from different data columns. Accuracy values higher than the average are</head><label></label><figDesc></figDesc><table><row><cell>url</cell><cell>description</cell><cell>extended</cell></row><row><cell></cell><cell></cell><cell>description</cell></row><row><cell>0.0439</cell><cell>0.1080</cell><cell>0.1250</cell></row><row><cell>0.0410</cell><cell>0.1194</cell><cell>0.1237</cell></row><row><cell>0.0318</cell><cell>0.0598</cell><cell>0.1406</cell></row><row><cell>0.0620</cell><cell>0.1911</cell><cell>0.2023</cell></row><row><cell>0.1560</cell><cell>0.0951</cell><cell>0.0322</cell></row><row><cell>0.3217</cell><cell>0.0877</cell><cell>0.0319</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_47"><head>Table 4 . The Post-Core dataset size</head><label>4</label><figDesc></figDesc><table><row><cell></cell><cell>bookmark</cell><cell>bibtex</cell><cell>tas</cell><cell># of users</cell></row><row><cell>Training</cell><cell>37037</cell><cell>17267</cell><cell>218682</cell><cell>982</cell></row><row><cell>Validation</cell><cell>4231</cell><cell>5585</cell><cell>34933</cell><cell>433</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_48"><head>Table 5 . The Cleaned Dump dataset size</head><label>5</label><figDesc></figDesc><table><row><cell></cell><cell>bookmark</cell><cell>bibtex</cell><cell>tas</cell><cell># of users</cell></row><row><cell>Training</cell><cell>212373</cell><cell>122115</cell><cell>1101387</cell><cell>2689</cell></row><row><cell>Validation</cell><cell>50631</cell><cell>36809</cell><cell>299717</cell><cell>1292</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_49"><head>Table 6 . The test dataset size</head><label>6</label><figDesc></figDesc><table><row><cell></cell><cell>bookmark</cell><cell>bibtex</cell><cell># of users</cell></row><row><cell>Cleaned Dump</cell><cell>16898</cell><cell>26104</cell><cell>1591</cell></row><row><cell>Core</cell><cell>431</cell><cell>347</cell><cell>136</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_50"><head>Table 7 . Final results on the test dataset (Post-Core, Task #1)</head><label>7</label><figDesc></figDesc><table><row><cell># of tags</cell><cell>Recall</cell><cell>Precision</cell><cell>F-measure</cell></row><row><cell>1</cell><cell>0.074695721</cell><cell>0.243523557</cell><cell>0.114324737</cell></row><row><cell>2</cell><cell>0.121408237</cell><cell>0.213594717</cell><cell>0.154817489</cell></row><row><cell>3</cell><cell>0.152896044</cell><cell>0.193533634</cell><cell>0.170831363</cell></row><row><cell>4</cell><cell>0.175617505</cell><cell>0.179512968</cell><cell>0.177543872</cell></row><row><cell>5</cell><cell>0.191311486</cell><cell>0.169508783</cell><cell>0.179751416</cell></row><row><cell>6</cell><cell>0.203439061</cell><cell>0.162068819</cell><cell>0.180412681</cell></row><row><cell>7</cell><cell>0.213460494</cell><cell>0.156249045</cell><cell>0.180428119</cell></row><row><cell>8</cell><cell>0.22072531</cell><cell>0.151207336</cell><cell>0.179469517</cell></row><row><cell>9</cell><cell>0.227309809</cell><cell>0.147113534</cell><cell>0.178623208</cell></row><row><cell>10</cell><cell>0.232596191</cell><cell>0.143564862</cell><cell>0.177544378</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_51"><head>Table 8 . Final results on the test dataset (Cleaned Dump, Task #2)</head><label>8</label><figDesc></figDesc><table><row><cell># of tags</cell><cell>Recall</cell><cell>Precision</cell><cell>F-measure</cell></row><row><cell>1</cell><cell>0.142522512</cell><cell>0.42159383</cell><cell>0.213029148</cell></row><row><cell>2</cell><cell>0.241682971</cell><cell>0.367609254</cell><cell>0.291633121</cell></row><row><cell>3</cell><cell>0.315328224</cell><cell>0.331191088</cell><cell>0.323065052</cell></row><row><cell>4</cell><cell>0.367734647</cell><cell>0.295308483</cell><cell>0.327565903</cell></row><row><cell>5</cell><cell>0.406172737</cell><cell>0.264524422</cell><cell>0.320390826</cell></row><row><cell>6</cell><cell>0.443927734</cell><cell>0.242502142</cell><cell>0.313661833</cell></row><row><cell>7</cell><cell>0.47018359</cell><cell>0.221477537</cell><cell>0.301115964</cell></row><row><cell>8</cell><cell>0.49385481</cell><cell>0.204859836</cell><cell>0.289591798</cell></row><row><cell>9</cell><cell>0.509440246</cell><cell>0.190310422</cell><cell>0.27710381</cell></row><row><cell>10</cell><cell>0.520841594</cell><cell>0.176357265</cell><cell>0.263494978</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_52"><head>Table 2 .</head><label>2</label><figDesc>Results of the ARKTiS system</figDesc><table><row><cell cols="2">#(tags) recall precision f-score</cell><cell cols="2">#(tags) recall precision f-score</cell></row><row><cell>1</cell><cell>0.0025 0.0114 0.0041</cell><cell>1</cell><cell>0.0305 0.1072 0.0475</cell></row><row><cell>2</cell><cell>0.0025 0.0057 0.0035</cell><cell>2</cell><cell>0.0595 0.1082 0.0768</cell></row><row><cell>3</cell><cell>0.0039 0.0057 0.0046</cell><cell>3</cell><cell>0.0839 0.1064 0.0938</cell></row><row><cell>4</cell><cell>0.0041 0.0046 0.0043</cell><cell>4</cell><cell>0.1032 0.1032 0.1032</cell></row><row><cell>5</cell><cell>0.0053 0.0058 0.0055</cell><cell>5</cell><cell>0.1179 0.0995 0.1079</cell></row><row><cell></cell><cell>Table 1. Baseline</cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_53"><head>Table 1 .</head><label>1</label><figDesc>Top terms composing the latent topic "images" and "tutorial"</figDesc><table><row><cell>Tag</cell><cell>Count Prob.</cell><cell>Tag</cell><cell>Count Prob.</cell></row><row><cell>images(tag)</cell><cell cols="2">243 0.064 tutorial(tag)</cell><cell>640 0.185</cell></row><row><cell>photo(tag)</cell><cell cols="2">218 0.057 howto(tag)</cell><cell>484 0.140</cell></row><row><cell>photography(tag)</cell><cell cols="2">205 0.054 tutorial(desc)</cell><cell>204 0.059</cell></row><row><cell>image(tag)</cell><cell cols="2">188 0.049 tutorials(tag)</cell><cell>184 0.053</cell></row><row><cell>photos(tag)</cell><cell cols="2">164 0.043 tutorials(desc)</cell><cell>173 0.050</cell></row><row><cell>photo(desc)</cell><cell>138 0.036</cell><cell>tips(tag)</cell><cell>126 0.037</cell></row><row><cell>images(desc)</cell><cell cols="2">106 0.028 reference(tag)</cell><cell>118 0.034</cell></row><row><cell>photos(desc)</cell><cell>98 0.026</cell><cell>guide(tag)</cell><cell>79 0.023</cell></row><row><cell>flickr(tag)</cell><cell cols="2">93 0.024 lessons(tag)</cell><cell>50 0.014</cell></row><row><cell>pictures(desc)</cell><cell>61 0.016</cell><cell>tips(desc)</cell><cell>48 0.014</cell></row><row><cell>graphics(tag)</cell><cell cols="2">49 0.013 wschools(desc)</cell><cell>45 0.013</cell></row><row><cell>media(tag)</cell><cell cols="2">48 0.013 tutoriel(tag)</cell><cell>33 0.010</cell></row><row><cell>art(tag)</cell><cell cols="2">48 0.013 comment(tag)</cell><cell>29 0.008</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_54"><head>Table 2 .</head><label>2</label><figDesc>Fields parsed to represent a resource</figDesc><table><row><cell cols="2">Bibtex</cell><cell>Bookmark</cell></row><row><cell>Author</cell><cell>Title</cell><cell>URL</cell></row><row><cell cols="3">Editor Description Description</cell></row><row><cell cols="2">Booktitle Journal</cell><cell>Extended</cell></row><row><cell>Abstract</cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_55"><head>Table 3 .</head><label>3</label><figDesc>Actual tags and recommended tags with computed probablity for URL http://jo.irisson.free.fr/bstdatabase/</figDesc><table><row><cell>Real Tag</cell><cell>LDA Tag</cell><cell>LDA Prob.</cell></row><row><cell>latex</cell><cell>bibtex(tag)</cell><cell>0.017</cell></row><row><cell>bibtex</cell><cell>latex(tag)</cell><cell>0.017</cell></row><row><cell>bibliography</cell><cell>bibtex(desc)</cell><cell>0.014</cell></row><row><cell>database</cell><cell>latex(desc)</cell><cell>0.008</cell></row><row><cell>engine</cell><cell>theory(desc)</cell><cell>0.005</cell></row><row><cell>style</cell><cell>citeulike(desc)</cell><cell>0.005</cell></row><row><cell>tex</cell><cell>bibliography(tag)</cell><cell>0.004</cell></row><row><cell>reference</cell><cell>database(tag)</cell><cell>0.003</cell></row><row><cell>academic</cell><cell>styles(desc)</cell><cell>0.003</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_56"><head>Table 4 .</head><label>4</label><figDesc>F-measure for different number of recommended tags and different number of LDA topics compared with recommending the most frequent tags (mf)</figDesc><table><row><cell>No. Tags</cell><cell># LDA topics 50 100 200 400 600 800 1000 2500 5000 10000 mf</cell></row><row><cell>1</cell><cell>0.170 0.191 0.214 0.229 0.229 0.230 0.229 0.238 0.240 0.235 0.270</cell></row><row><cell>2</cell><cell>0.200 0.225 0.248 0.266 0.271 0.271 0.274 0.289 0.288 0.283 0.335</cell></row><row><cell>3</cell><cell>0.209 0.233 0.257 0.277 0.282 0.285 0.287 0.302 0.303 0.300 0.362</cell></row><row><cell>4</cell><cell>0.209 0.237 0.257 0.279 0.287 0.289 0.292 0.305 0.307 0.303 0.379</cell></row><row><cell>5</cell><cell>0.209 0.238 0.258 0.280 0.286 0.291 0.293 0.307 0.307 0.304 0.388</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_57"><head>Table 5 .</head><label>5</label><figDesc>Evaluation results for tag recommendation based on most frequent tags, based on 5000 latent topics, and their combination with λ = 0.5.</figDesc><table><row><cell>No. Tags</cell><cell cols="3">Most Frequent Tags Recall Prec. F-Meas. Recall Prec. F-Meas. Recall Prec. F-Meas. Latent Topics Combination</cell></row><row><cell>1</cell><cell>0.190 0.467 0.270</cell><cell>0.165 0.437 0.240</cell><cell>0.214 0.537 0.306</cell></row><row><cell>2</cell><cell>0.274 0.430 0.335</cell><cell>0.232 0.380 0.288</cell><cell>0.302 0.479 0.370</cell></row><row><cell>3</cell><cell>0.329 0.403 0.362</cell><cell>0.271 0.343 0.302</cell><cell>0.357 0.441 0.394</cell></row><row><cell>4</cell><cell>0.370 0.388 0.379</cell><cell>0.298 0.316 0.307</cell><cell>0.393 0.415 0.404</cell></row><row><cell>5</cell><cell>0.400 0.377 0.388</cell><cell>0.316 0.299 0.307</cell><cell>0.421 0.398 0.409</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_58"><head>Table 6 .</head><label>6</label><figDesc>Evaluation results DC09 challenge Task 1 based on most frequent tags, based on 5000 latent topics, and their combination with λ = 0.5. Setups and Results for the Challenge Submission We have submitted tag recommendations for Task 1 and Task 2 in the ECML PKDD Discovery Challenge 2009. Task 1 aims at recommending tags for arbitrary users annotating a resource in 2009 based on tag assignments until 2008. Thus the test data contain tags, resources, and users which are not available in the training data. The topic models have been trained on the full dataset, comprising about 9.3 Mio tokens for 415 K resources. The test set consists of 43002 posts. Table</figDesc><table><row><cell>No. Tags</cell><cell cols="3">Most Frequent Tags Recall Prec. F-Meas. Recall Prec. F-Meas. Recall Prec. F-Meas. Latent Topics Combination</cell></row><row><cell>1</cell><cell>0.010 0.032 0.015</cell><cell>0.045 0.158 0.070</cell><cell>0.049 0.169 0.076</cell></row><row><cell>2</cell><cell>0.018 0.031 0.022</cell><cell>0.073 0.131 0.094</cell><cell>0.078 0.140 0.100</cell></row><row><cell>3</cell><cell>0.022 0.029 0.025</cell><cell>0.092 0.114 0.102</cell><cell>0.099 0.122 0.110</cell></row><row><cell>4</cell><cell>0.026 0.028 0.027</cell><cell>0.094 0.112 0.103</cell><cell>0.102 0.120 0.111</cell></row><row><cell>5</cell><cell>0.028 0.027 0.028</cell><cell>0.096 0.112 0.103</cell><cell>0.105 0.120 0.112</cell></row><row><cell>3.3</cell><cell></cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_59"><head>Table 7</head><label>7</label><figDesc>again compares the results</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_60"><head>Table 7 .</head><label>7</label><figDesc>Evaluation results DC09 challenge Task 2 based on most frequent tags, based on 5000 latent topics, and their combination with λ = 0.5.</figDesc><table><row><cell>No. Tags</cell><cell cols="3">Most Frequent Tags Recall Prec. F-Meas. Recall Prec. F-Meas. Recall Prec. F-Meas. Latent Topics Combination</cell></row><row><cell>1</cell><cell>0.147 0.411 0.216</cell><cell>0.133 0.404 0.200</cell><cell>0.156 0.450 0.232</cell></row><row><cell>2</cell><cell>0.223 0.341 0.270</cell><cell>0.204 0.326 0.251</cell><cell>0.252 0.386 0.305</cell></row><row><cell>3</cell><cell>0.284 0.305 0.294</cell><cell>0.258 0.281 0.269</cell><cell>0.313 0.339 0.326</cell></row><row><cell>4</cell><cell>0.325 0.275 0.298</cell><cell>0.298 0.251 0.272</cell><cell>0.352 0.300 0.324</cell></row><row><cell>5</cell><cell>0.357 0.256 0.298</cell><cell>0.319 0.224 0.263</cell><cell>0.386 0.276 0.322</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_61"><head>Table 8 .</head><label>8</label><figDesc>Evaluation results for DC09 challenge Task 2 for 5000 latent topics without content</figDesc><table><row><cell cols="4">No. Tags Recall Precision F-Measure</cell></row><row><cell>1</cell><cell>0.128</cell><cell>0.362</cell><cell>0.189</cell></row><row><cell>2</cell><cell>0.191</cell><cell>0.293</cell><cell>0.232</cell></row><row><cell>3</cell><cell>0.236</cell><cell>0.254</cell><cell>0.245</cell></row><row><cell>4</cell><cell>0.267</cell><cell>0.225</cell><cell>0.244</cell></row><row><cell>5</cell><cell>0.299</cell><cell>0.207</cell><cell>0.245</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_62"><head>Table 1 .</head><label>1</label><figDesc>Candidate Set based Tag Recommendation AlgorithmInput: testing sample: T j = {D j , U j }, threshold N and L Output: top N tags t = {t1, ..., tN } ∈ CS with P (t k |D j ) in descending order 10. return top N tags in C as t</figDesc><table><row><cell cols="2">1. candidate set CS ← ∅</cell></row><row><cell cols="2">2. for w in D j</cell></row><row><cell>3.</cell><cell>add w into CS</cell></row><row><cell>4.</cell><cell>add top L tags t into CS according to P (t|w)</cell></row><row><cell cols="2">5. end for</cell></row><row><cell cols="2">6. for each word t k ∈ C</cell></row><row><cell>7.</cell><cell>compute P (t k |D j ) using (9)</cell></row><row><cell cols="2">8. end for</cell></row><row><cell cols="2">9. sort t k</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_63"><head>Table 2 .</head><label>2</label><figDesc>Statistics of Cleaned Dump &amp; Post Core datasets</figDesc><table><row><cell></cell><cell>tag assignments</cell><cell>number of posts</cell><cell>number of users</cell></row><row><cell>Cleaned Dump</cell><cell cols="2">1,401,104 263, 004 / 158, 924 : 421, 928</cell><cell>3, 617</cell></row><row><cell>Post Core</cell><cell>253,615</cell><cell>41,268 / 22,852 : 64, 120</cell><cell>1, 185</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_64"><head>Table 3 .</head><label>3</label><figDesc>Fields of Three Dataset Tables</figDesc><table><row><cell>table</cell><cell>fields</cell></row><row><cell>tas</cell><cell>user, tag, content type, content id, date</cell></row><row><cell cols="2">bookmark content id, url hash, URL, description, extended description, date</cell></row><row><cell>bibtex</cell><cell>content id, journal, chapter, edition, month, day, booktitle,</cell></row><row><cell></cell><cell>howPublished, institution, organization, publisher, address, school,</cell></row><row><cell></cell><cell>series, bibtexKey, url, type, description, annote, note, pages, bKey,</cell></row><row><cell></cell><cell>number, crossref, misc, bibtexAbstract, simhash0, simhash1, simhash2,</cell></row><row><cell></cell><cell>entrytype, title, author, editor, year</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_65"><head>Table 5 .</head><label>5</label><figDesc>Top 10 Tags and their Frequency Cleaned Dump bookmarks:52795 → zzztosort:11839 → video:10788→</figDesc><table /><note>software:10171 → programming:9491 → indexforum:9183 → web20:8777 → books:7934 → media:7149 → tools:6903 Post Core web20:4474 → software:3867 → juergen:3092 → tools:3058 → web:2930 → tagging:2196 → semanticweb:2055 → folksonomy:1944 → search:1896 → bookmarks:1840 6 Experimental Result 6.1 Tagging Performance</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_66"><head>Table 6 .</head><label>6</label><figDesc>Sample Domains with Top 5 used tags</figDesc><table><row><cell>domain</cell><cell>tags and their previously used probability</cell></row><row><cell>www.apple.com</cell><cell>apple:0.17 mac:0.13 software:0.09 osx:0.07 bookmarks:0.07</cell></row><row><cell>answers.yahoo.com</cell><cell>knowledge:0.14 yahoo:0.14 web20:0.07 all:0.07 answer:0.07</cell></row><row><cell>ant.apache.org</cell><cell>java:0.19 ant:0.17 programming:0.07 apache:0.07 tool:0.07</cell></row><row><cell>picasa.google.com</cell><cell>google:0.21 image:0.14 download:0.14 linux:0.14 picasa:0.14</cell></row><row><cell cols="2">research.microsoft.com microsoft:0.10 research:0.09 people:0.04 social:0.04 award:0.03</cell></row><row><cell cols="2">www.research.ibm.com ibm:0.11 datamining:0.07 software:0.04 machinelearning:0.04</cell></row><row><cell></cell><cell>journal:0.04</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_67"><head>Table 7 .</head><label>7</label><figDesc>Performance for Task 1 ( α = 0.15, λ = 0.05, β = 0.05, γ = 0.5, θ = 0.25 for bookmark, α = 0.15, λ = 0.05, β = 0.1, γ = 0.7 for bibtex with P 2</figDesc><table><row><cell>tr )</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_68"><head>Table 8 .</head><label>8</label><figDesc>Performance</figDesc><table><row><cell>tr )</cell></row></table><note>on Task 2 data( α = 0.15, λ = 0.05, β = 0.05, γ = 0.5, θ = 0.25 for bookmark, α = 0.15, λ = 0.05, β = 0.1, γ = 0.7 for bibtex with P 2</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_70"><head>Table 10 .</head><label>10</label><figDesc>Sampled Words with their top tags ti: P 1</figDesc><table><row><cell>tr (ti|w)(EM); P 2 tr (CO)</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_71"><head>Table 1 .</head><label>1</label><figDesc>Statistics</figDesc><table><row><cell></cell><cell>BibSonomy</cell><cell>CiteULike</cell></row><row><cell>number of tags</cell><cell>1,401,104</cell><cell>4,927,383</cell></row><row><cell>number of unique tags</cell><cell cols="2">93,756 (7% of tags) 206,911 (4% of tags)</cell></row><row><cell>number of bookmark posts</cell><cell>263,004</cell><cell>N/A</cell></row><row><cell>number of unique bookmarks</cell><cell>235,328 (89% of posts)</cell><cell>N/A</cell></row><row><cell>number of BibTeX posts</cell><cell>158,924</cell><cell>1,610,011</cell></row><row><cell cols="3">number of unique BibTeX entries 143,050 (90% of posts) 1,390,747 (86% of posts)</cell></row><row><cell>number of users</cell><cell>3,617 (1% of posts)</cell><cell>42,452 (3% of posts)</cell></row></table><note>of BibSonomy training data compared to CiteULike dataset (complete dump up to February 27, 2009). Both datasets have similar proportion of unique tags, posts and users.</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_72"><head>Table</head><label></label><figDesc></figDesc><table><row><cell>Posts: Social tag prediction</cell><cell cols="2">Towards the Semantic Web:</cell></row><row><cell></cell><cell cols="2">Collaborative Tag Suggestions</cell></row><row><cell>User A Heymann 08 tag recommendation</cell><cell cols="2">Xu 06 tag recommendation</cell></row><row><cell cols="3">User B prediction tag recommender social tagging tag recommender tagging</cell></row><row><cell>User C folksonomy prediction recommender social</cell><cell>folksonomy</cell><cell>recommender</cell></row><row><cell>tag tagging toread</cell><cell cols="2">summerschool tagging</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_73"><head></head><label></label><figDesc>Scontent ←− mergeSumP rob(S title , SURL) /*Step 2 -Retrieval of resource related tags*/ S T itleT oT ag ←− ∅// related tags from TitleToTag graph ST agT oT ag ←− ∅// related tags from TagToTag graph S Pr ←− getP rof ileRecommendationBasic(Pr) foreach s k ∈ S title do S s</figDesc><table><row><cell>Algorithm 1: Tag recommendation system</cell></row><row><cell>Data: a resource r and user u</cell></row><row><cell>Result: a tag recommendation set S f inal</cell></row><row><cell>begin</cell></row><row><cell>/*Step 1 -Extraction of content based tags*/</cell></row><row><cell>W ords title ←− extractT itleW ords(r)</cell></row><row><cell>S title ←− ∅</cell></row><row><cell>foreach w ∈ W ords title do</cell></row><row><cell>S title add makeT ag(w, getP riorU sef ullness(w))</cell></row><row><cell>removeLowQualityT ags(S title , 0.05)</cell></row><row><cell>if isBookmark(r) then</cell></row><row><cell>W ordsURL ←− extractU rlW ords(r)</cell></row><row><cell>SURL ←− ∅</cell></row><row><cell>foreach w ∈ W ordsURL do</cell></row><row><cell>SURL add makeT ag(w, getP riorU sef ullness(w))</cell></row><row><cell>removeLowQualityT ags(SURL, 0.05)</cell></row><row><cell>rescoreLeadingP recision(S title , 0.2)</cell></row><row><cell>rescoreLeadingP recision(SURL, 0.1)</cell></row><row><cell>/*Final recommendation*/</cell></row><row><cell>rescoreLeadingP recision(S title , 0.3)</cell></row><row><cell>rescoreLeadingP recision(S Pr , 0.3)</cell></row><row><cell>rescoreLeadingP recision(S r,u Related , 0.45)</cell></row><row><cell>S f inal ←− unionP rob(S title , S Pr , S r,u Related )</cell></row><row><cell>end</cell></row></table><note>k ,T itleT oT ag ←− ∅ foreach t ∈ getRelated(g T itleT oT ag , s k ) do S s k ,T itleT oT ag add makeT ag(t, s k .l * conf idenceT itleT oT ag(s k .t, t)) foreach s k ∈ Scontent do Ss k ,T agT oT ag ←− ∅ foreach t ∈ getRelated(gT agT oT ag , s k ) do Ss k ,T agT oT ag add makeT ag(t, s k .l * conf idenceT agT oT ag(s k .t, t)) S T itleT oT ag ←− unionP rob(T s 1 ,T itleT oT ag , . . . , T sn,T itleT oT ag ) ST agT oT ag ←− unionP rob(Ts 1 ,T agT oT ag , . . . , Ts n,T agT oT ag ) S r Related ←− unionP rob(S T itleT oT ag , ST agT oT ag , S Pr ) /*Step 3 -Retrieval of resource and user related tags*/ S Pu ←− getP rof ileRecommendationByDay(Pu) S r,u Related ←− indersectionP rob(S r Related , S Pu )</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_75"><head>Table 3 .</head><label>3</label><figDesc>Number of posts in training and test dataset.</figDesc><table><row><cell>Sparsity of folksonomy</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_77"><head></head><label></label><figDesc>System based on Logistic RegressionE. Montañés 1 , J. R. Quevedo 2 , I. Díaz 1 , and J. Ranilla1  {montaneselena,quevedo,sirene,ranilla}@uniovi.es</figDesc><table><row><cell>1 Computer Science Department, University of Oviedo (Spain)</cell></row><row><cell>2 Artificial Intelligence Center, University of Oviedo (Spain)</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_78"><head></head><label></label><figDesc>)Let p 7 = (u 2 , r 2 , {t 2 , t 5 }) be a randomly selected test post at instant d 7 . Therefore the test set is formed by</figDesc><table><row><cell cols="3">Example 1 Let the following folksonomy be</cell><cell></cell></row><row><cell cols="4">post date U ser Resource T ags</cell></row><row><cell cols="2">p 1 d 1 u 1</cell><cell>r 1</cell><cell>t 1</cell></row><row><cell cols="2">p 2 d 2 u 1</cell><cell>r 2</cell><cell>t 2</cell></row><row><cell cols="2">p 3 d 3 u 2</cell><cell>r 1</cell><cell>t 1</cell></row><row><cell cols="2">p 4 d 4 u 3</cell><cell>r 1</cell><cell>t 3</cell><cell>(2)</cell></row><row><cell cols="2">p 5 d 5 u 2</cell><cell>r 2</cell><cell>t 4</cell></row><row><cell cols="2">p 6 d 6 u 2</cell><cell>r 1</cell><cell>t 2 , t 3</cell></row><row><cell cols="2">p 7 d 7 u 2</cell><cell>r 2</cell><cell>t 2 , t 5</cell></row><row><cell cols="2">p 8 d 8 u 3</cell><cell>r 2</cell><cell>t 1</cell></row><row><cell cols="4">example date U ser Resource T ags</cell></row><row><cell>e 1</cell><cell>d 7 u 2</cell><cell>r 2</cell><cell>t 2</cell></row><row><cell>e 2</cell><cell>d 7 u 2</cell><cell>r 2</cell><cell>t 5</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_79"><head></head><label></label><figDesc>{t i1 , ..., t in })} /p j = (u i , r i , {t i1 , ..., t in })} = {{p 3 , p 5 , p 6 } ∪ {p 2 , p 5 }}\{p 5 } = {p 2 , p 3 , p 6 }Therefore the training set is defined as follows.</figDesc><table><row><cell cols="4">Example 2 Let us show an example of each training set for the test set of Example</cell></row><row><cell>1.</cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="4">In this case the training set is computed as follows.</cell></row><row><cell></cell><cell>U R d7 u2,r2 =</cell><cell></cell><cell></cell></row><row><cell cols="4">{P d7 u2 ∪ R d7 r2 }\{p j example date U ser Resource T ags</cell></row><row><cell>e 2</cell><cell>d 2 u 1</cell><cell>r 2</cell><cell>t 2</cell></row><row><cell>e 3</cell><cell>d 3 u 2</cell><cell>r 1</cell><cell>t 1</cell></row><row><cell>e 61</cell><cell>d 6 u 2</cell><cell>r 1</cell><cell>t 2</cell></row><row><cell>e 62</cell><cell>d 6 u 2</cell><cell>r 1</cell><cell>t 3</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_80"><head></head><label></label><figDesc>Thus, this is another advantage of building a training set particularly for each test post. Let us consider the test post of Example 1 and the training set of Example 2. The features for the test post are t 2 and t 4 , hence, the training set of Approach 3 in Example 2 will be reduced to be represented at most with these two tags. Originally, that training set has the following representation:</figDesc><table><row><cell cols="5">Example 4 example date resource f eatures category</cell></row><row><cell>e 2</cell><cell>d 2</cell><cell>r 2</cell><cell>∅</cell><cell>t 2</cell></row><row><cell>e 3</cell><cell>d 3</cell><cell>r 1</cell><cell>t 1</cell><cell>t 1</cell></row><row><cell>e 61</cell><cell>d 6</cell><cell>r 1</cell><cell>t 1 , t 3</cell><cell>t 2</cell></row><row><cell>e 62</cell><cell>d 6</cell><cell>r 1</cell><cell>t 1 , t 2 , t 3</cell><cell>t 3</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_82"><head>Table 1 .</head><label>1</label><figDesc>The parameters and performance of the best settings in training data for all post collections Table1shows the best setting parameters together with their F 1 computed when at most 5 tags are returned, obtained using training sets. The parameters of Table1were established to classify the test datasets, obtaining the results shown in Table2.</figDesc><table><row><cell>Data</cell><cell cols="4">Kind of Collaboration Measure Penalizing degree F1</cell></row><row><cell>'bm09 no core'</cell><cell>Intersection</cell><cell>ILir</cell><cell>0</cell><cell>28.54%</cell></row><row><cell>'bt09 no core'</cell><cell>Intersection</cell><cell>ILir</cell><cell>0.0625</cell><cell>28.56%</cell></row><row><cell>'bm09 core'</cell><cell>Intersection</cell><cell>IG</cell><cell>0</cell><cell>30.90%</cell></row><row><cell>'bt09 core'</cell><cell>Intersection</cell><cell>IG</cell><cell>0.0625</cell><cell>37.07%</cell></row><row><cell></cell><cell>Data</cell><cell></cell><cell>F1</cell><cell></cell></row><row><cell></cell><cell>'bm09 no core'</cell><cell></cell><cell>7.28%</cell><cell></cell></row><row><cell></cell><cell>'bt09 no core'</cell><cell></cell><cell>6.75%</cell><cell></cell></row><row><cell></cell><cell>'bm09 core'</cell><cell></cell><cell>24.21%</cell><cell></cell></row><row><cell></cell><cell>'bt09 core'</cell><cell></cell><cell>28.76%</cell><cell></cell></row><row><cell></cell><cell cols="3">Task 1 'bm09 and bt09 no core' 6.98%</cell><cell></cell></row><row><cell></cell><cell cols="2">Task 2 'bm09 and bt09 core'</cell><cell>26.25%</cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_83"><head>Table 2 .</head><label>2</label><figDesc>The performance of test data for all post collections</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_84"><head>Table 3 .</head><label>3</label><figDesc>The effect of collaboration, post selection and a penalization of the oldest posts</figDesc><table><row><cell>Data</cell><cell cols="3">F1 no Collaboration F1 no post selection F1 no penalization</cell></row><row><cell>'bm09 no core'</cell><cell>28.02%</cell><cell>27.25%</cell><cell>28.54%</cell></row><row><cell>'bt09 no core'</cell><cell>28.53%</cell><cell>27.30%</cell><cell>28.51%</cell></row><row><cell>'bm09 core'</cell><cell>29.32%</cell><cell>26.32%</cell><cell>30.90%</cell></row><row><cell>'bt09 core'</cell><cell>36.10%</cell><cell>34.74%</cell><cell>36.84%</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_85"><head>Table 1 .</head><label>1</label><figDesc>Comparison of the results for test and training data set</figDesc><table><row><cell cols="5">Dataset Recommender Supplied Recall Precision F1-Score</cell></row><row><cell></cell><cell></cell><cell>Posts</cell><cell></cell><cell></cell></row><row><cell cols="2">Training citeulike.org</cell><cell>5285 0.446</cell><cell>0.134</cell><cell>0.206</cell></row><row><cell></cell><cell></cell><cell>/ 6372</cell><cell></cell><cell></cell></row><row><cell></cell><cell>del.icio.us</cell><cell>38383 0.418</cell><cell>0.353</cell><cell>0.383</cell></row><row><cell></cell><cell></cell><cell>/ 40882</cell><cell></cell><cell></cell></row><row><cell></cell><cell>tagthe.net</cell><cell>40468 0.066</cell><cell>0.053</cell><cell>0.059</cell></row><row><cell></cell><cell></cell><cell>/ 51580</cell><cell></cell><cell></cell></row><row><cell></cell><cell>bibtex</cell><cell>22341 0.155</cell><cell>0.127</cell><cell>0.139</cell></row><row><cell></cell><cell></cell><cell>/ 22341</cell><cell></cell><cell></cell></row><row><cell></cell><cell>web content</cell><cell>33193 0.150</cell><cell>0.123</cell><cell>0.135</cell></row><row><cell></cell><cell></cell><cell>/ 40882</cell><cell></cell><cell></cell></row><row><cell></cell><cell>meta</cell><cell>63107 0.350</cell><cell>0.254</cell><cell>0.294</cell></row><row><cell></cell><cell></cell><cell>/ 64120</cell><cell></cell><cell></cell></row><row><cell></cell><cell>meta</cell><cell>62104 0.344</cell><cell>0.269</cell><cell>0.302</cell></row><row><cell></cell><cell>with filter</cell><cell>/ 64120</cell><cell></cell><cell></cell></row><row><cell>Test</cell><cell>Meta/Overall</cell><cell>30844 0.132</cell><cell>0.103</cell><cell>0.116</cell></row><row><cell></cell><cell>with filter</cell><cell>/ 43002</cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_86"><head>Table 2 .</head><label>2</label><figDesc>Comparison of the results for test and training data set</figDesc><table><row><cell>Dataset</cell><cell></cell><cell cols="3">Recommender Recall Precision F1-Score</cell></row><row><cell cols="2">Training Bookmarks</cell><cell>del.icio.us</cell><cell>0.391 0.280</cell><cell>0.326</cell></row><row><cell></cell><cell></cell><cell cols="2">WebCrawler 0.139 0.099</cell><cell>0.116</cell></row><row><cell></cell><cell></cell><cell>DataSet</cell><cell>0.113 0.092</cell><cell>0.102</cell></row><row><cell></cell><cell>Bibtex</cell><cell cols="2">WebCrawler 0.250 0.128</cell><cell>0.169</cell></row><row><cell></cell><cell></cell><cell cols="2">GoogleScholar 0.087 0.073</cell><cell>0.079</cell></row><row><cell></cell><cell></cell><cell>DataSet</cell><cell>0.083 0.061</cell><cell>0.070</cell></row><row><cell></cell><cell>Meta</cell><cell></cell><cell>0.334 0.213</cell><cell>0.260</cell></row><row><cell>Test</cell><cell>Meta/Overall</cell><cell></cell><cell>0.214 0.155</cell><cell>0.180</cell></row><row><cell></cell><cell>Meta</cell><cell cols="2">w CiteULike 0.218 0.157</cell><cell>0.183</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_87"><head>Table 3 .</head><label>3</label><figDesc>Additional quantitative data</figDesc><table><row><cell></cell><cell>Competition Training</cell></row><row><cell>Tags harvested</cell><cell>1432413 8389379</cell></row><row><cell>Bookmarks tagged</cell><cell>16898 263004</cell></row><row><cell>Bibtex entries tagged</cell><cell>26104 158924</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_88"><head>Table 4 .</head><label>4</label><figDesc>Comparison of the results for test and training data set</figDesc><table><row><cell cols="5">Dataset Recommender Supplied Recall Precision F1-Score</cell></row><row><cell></cell><cell></cell><cell>Posts</cell><cell></cell><cell></cell></row><row><cell cols="2">Training 1. by Resource</cell><cell>64120 0.731</cell><cell>0.586</cell><cell>0.651</cell></row><row><cell></cell><cell>2. by User</cell><cell>64120 0.349</cell><cell>0.205</cell><cell>0.258</cell></row><row><cell></cell><cell>3. by User-Sim.</cell><cell>34738 0.265</cell><cell>0.271</cell><cell>0.268</cell></row><row><cell></cell><cell>1. + 2. + 3</cell><cell>64120 0.774</cell><cell>0.517</cell><cell>0.620</cell></row><row><cell></cell><cell>1. + 2.</cell><cell>64120 0.846</cell><cell>0.570</cell><cell>0.681</cell></row><row><cell></cell><cell>1. + 2.</cell><cell>64120 0.846</cell><cell>0.576</cell><cell>0.685</cell></row><row><cell></cell><cell>with filter</cell><cell></cell><cell></cell><cell></cell></row><row><cell>Test</cell><cell>1. + 2. + 3.</cell><cell>778 0.389</cell><cell>0.262</cell><cell>0.313</cell></row><row><cell></cell><cell>with filter</cell><cell>/ 778</cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_90"><head>Table 1 .</head><label>1</label><figDesc>Statistics of Experiment data</figDesc><table><row><cell></cell><cell>Bookmark Bibtex</cell></row><row><cell>Training</cell><cell>181,491 72,124</cell></row><row><cell>Testing</cell><cell>16,898 26,104</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_91"><head>Table 2 .</head><label>2</label><figDesc>Performance of two-level learning hierarchy of multi-concept based keyword extraction method for each number of recommended tags using the optimal parameter α = 0.05</figDesc><table><row><cell cols="3">Num. of Tags Recall Precision F-Measure</cell></row><row><cell>1</cell><cell>0.0538 0.1832</cell><cell>0.0832</cell></row><row><cell>2</cell><cell>0.0908 0.1621</cell><cell>0.1164</cell></row><row><cell>3</cell><cell>0.1187 0.1498</cell><cell>0.1324</cell></row><row><cell>4</cell><cell>0.1386 0.1406</cell><cell>0.1396</cell></row><row><cell>5</cell><cell>0.1533 0.1345</cell><cell>0.1433</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_92"><head>Table 1 .</head><label>1</label><figDesc>Results of the ECML-PKDD 2009 Discovery Challenge</figDesc><table><row><cell cols="3">#Tag Precision Recall F1</cell></row><row><cell>1</cell><cell>19.51</cell><cell>6.89 10.19</cell></row><row><cell>2</cell><cell>16.34</cell><cell>10.10 12.53</cell></row><row><cell>3</cell><cell>14.55</cell><cell>12.16 13.25</cell></row><row><cell>4</cell><cell>13.56</cell><cell>13.53 13.55</cell></row><row><cell>5</cell><cell>13.56</cell><cell>13.53 13.55</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_93"><head>Table 1 .</head><label>1</label><figDesc>Tags ordered by number of uses and "popularity"</figDesc><table><row><cell cols="2">Number of uses Popularity</cell></row><row><cell>bookmarks</cell><cell>software</cell></row><row><cell>zzztosort</cell><cell>web</cell></row><row><cell>video</cell><cell>web20</cell></row><row><cell>software</cell><cell>video</cell></row><row><cell>programming</cell><cell>blog</cell></row><row><cell>web20</cell><cell>bookmarks</cell></row><row><cell>books</cell><cell>programming</cell></row><row><cell>media</cell><cell>internet</cell></row><row><cell>tools</cell><cell>tools</cell></row><row><cell>web</cell><cell>social</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_94"><head>Table 2 .</head><label>2</label><figDesc>Recommendation methods</figDesc><table><row><cell>Method</cell></row><row><cell>Collaborative filtering (UR neighbourhood)</cell></row><row><cell>Collaborative filtering (UT neighbourhood)</cell></row><row><cell>Most frequent tags by resource</cell></row><row><cell>Most frequent tags by resource (popularity &gt; 3)</cell></row><row><cell>Most frequent user tags</cell></row><row><cell>Most frequent user tags (popularity &gt; 3)</cell></row><row><cell>Most popular global tags</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_95"><head>Table 3 .</head><label>3</label><figDesc>Results on the competition set</figDesc><table><row><cell>Method</cell><cell>F-measure with 5 tags</cell></row><row><cell>CF-UR</cell><cell>0.2084</cell></row><row><cell>CF-UT</cell><cell>0.2317</cell></row><row><cell>resource tags</cell><cell>0.3067</cell></row><row><cell>resource tags (popularity &gt; 3)</cell><cell>0.2940</cell></row><row><cell>user tags</cell><cell>0.0935</cell></row><row><cell>user tags (popularity &gt; 3)</cell><cell>0.0050</cell></row><row><cell>popular tags</cell><cell>0.0354</cell></row><row><cell>combined</cell><cell>0.2952</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_98"><head>Table 1 .</head><label>1</label><figDesc>Basic statistics of the full bibtex and bookmark data sets. Mean Len. is the mean number of words in the corresponding text content.</figDesc><table><row><cell>Name</cell><cell cols="4">#posts #tags #users #words Mean Length Mean #tags/user</cell></row><row><cell>bibtex</cell><cell cols="2">158,912 50,855 1,790 278,106</cell><cell>47.67</cell><cell>60.75</cell></row><row><cell>bookmark</cell><cell cols="2">263,004 56,424 2,679 293,026</cell><cell>11.83</cell><cell>57.78</cell></row><row><cell>bibtex(pcore2)</cell><cell>22,852 5,816</cell><cell>788 48,401</cell><cell>59.21</cell><cell>31.75</cell></row><row><cell cols="2">bookmark(pcore2) 41,268 10,702</cell><cell>861 47,689</cell><cell>12.23</cell><cell>60.26</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_99"><head>Table 2 .</head><label>2</label><figDesc>P, R and F1 of search-based kNN and different learning methods for FDT on the bibtex dataset. All averaged over 5 folds.</figDesc><table><row><cell>Method</cell><cell>Precision</cell><cell>Recall</cell><cell>F1</cell></row><row><cell>search-based kNN</cell><cell>0.2792</cell><cell>0.2324</cell><cell>0.2537</cell></row><row><cell>FDT(TFIDF+CC)</cell><cell>0.2517</cell><cell>0.2152</cell><cell>0.2320</cell></row><row><cell>FDT(TFIDF+MI)</cell><cell>0.1822</cell><cell>0.1652</cell><cell>0.1733</cell></row><row><cell>FDT(TFIDF+χ 2 )</cell><cell>0.2261</cell><cell>0.2235</cell><cell>0.2248</cell></row><row><cell>FDT(TFITF+CC)</cell><cell>0.2513</cell><cell>0.2173</cell><cell>0.2330</cell></row><row><cell>FDT(TFITF+MI)</cell><cell>0.2432</cell><cell>0.2526</cell><cell>0.2478</cell></row><row><cell>FDT(TFITF+χ 2 )</cell><cell>0.2216</cell><cell>0.2246</cell><cell>0.2231</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_100"><head>Table 4 .</head><label>4</label><figDesc>P, R and F1 of search-based kNN and FDT on different set of users on the bibtex dataset. Trained(ALL) means that we train the model using posts from all users, and test the performance on given set of users. Train(USER) means that both training and testing use posts from the given set of users. The % column indicates the size of corresponding user group, as the percentage in all posts. All averaged over 5 folds.</figDesc><table><row><cell>Trained(ALL)</cell><cell>P</cell><cell>R</cell><cell>F1 Trained(USER)</cell><cell>P</cell><cell>R</cell><cell>F1</cell><cell>%</cell></row><row><cell>kNN(2463)</cell><cell cols="3">0.1323 0.1352 0.1337 kNN(2463)</cell><cell cols="4">0.1376 0.1377 0.1376 24.27</cell></row><row><cell>kNN(2651)</cell><cell cols="3">0.1561 0.1573 0.1567 kNN(2651)</cell><cell cols="4">0.1953 0.1910 0.1932 12.40</cell></row><row><cell>kNN(3180)</cell><cell cols="3">0.4278 0.2771 0.3364 kNN(3180)</cell><cell cols="4">0.4440 0.2807 0.3440 9.20</cell></row><row><cell>kNN(2732)</cell><cell cols="3">0.6267 0.3915 0.4819 kNN(2732)</cell><cell cols="4">0.6517 0.4422 0.5269 3.78</cell></row><row><cell>kNN(rest)</cell><cell cols="3">0.3207 0.2530 0.2829 kNN(rest)</cell><cell cols="4">0.3202 0.2579 0.2857 50.35</cell></row><row><cell>FDT(2463)</cell><cell cols="3">0.1066 0.1869 0.1358 FDT(2463)</cell><cell cols="4">0.1100 0.2055 0.1429 24.27</cell></row><row><cell>FDT(2651)</cell><cell cols="3">0.1022 0.1285 0.1138 FDT(2651)</cell><cell cols="4">0.1126 0.1818 0.1391 12.40</cell></row><row><cell>FDT(3180)</cell><cell cols="3">0.3656 0.3334 0.3488 FDT(3180)</cell><cell cols="4">0.3688 0.3274 0.3469 9.20</cell></row><row><cell>FDT(2732)</cell><cell cols="3">0.8763 0.4927 0.6308 FDT(2732)</cell><cell cols="4">0.3142 0.1814 0.2300 3.78</cell></row><row><cell>FDT(rest)</cell><cell cols="3">0.3101 0.2516 0.2778 FDT(rest)</cell><cell cols="4">0.3260 0.2559 0.2867 50.35</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_101"><head>Table 5 .</head><label>5</label><figDesc>P, R and F1 of search-based kNN and FDT on different set of users on the bookmark dataset. Trained(ALL) means that we train the model using posts from all users, and test the performance on given set of users. Train(USER) means that both training and testing use posts from the given set of users. The % column indicates the size of corresponding user group, as the percentage in all posts. All averaged over 5 folds.For final test, we use FDT(ITF+MI) for bibtex and FDT(IDF+CC) for bookmark. The test data of DC09 has a different distribution with the training data. Most top ranked users don't appear in the test data. So we removed the top ranked users from the training data, use the rest group of users to train the model for final suggestion. The p/r/f1 on final test data are 0.1388/0.1049/0.1189 respectively. Compared to the cross validation results, the performance dropped a lot on final test data. One reason is that FDT does not suggest tags that are not in the training data. There are 93756 tags in the training data and 34051 tags in the test data, the overlapped tags are only 15194. To achieve better performance, suggesting new tags should be considered in the future.</figDesc><table><row><cell>Trained(ALL)</cell><cell>P</cell><cell>R</cell><cell>F1 Trained(USER)</cell><cell>P</cell><cell>R</cell><cell>F1</cell><cell>%</cell></row><row><cell>kNN(1747)</cell><cell cols="3">0.5523 0.4877 0.5180 kNN(1747)</cell><cell cols="4">0.6513 0.5566 0.6003 19.90</cell></row><row><cell>kNN(2977)</cell><cell cols="3">0.4554 0.4072 0.4299 kNN(2977)</cell><cell cols="4">0.5567 0.5154 0.5353 9.48</cell></row><row><cell>kNN(483)</cell><cell cols="3">0.1002 0.1365 0.1156 kNN(483)</cell><cell cols="4">0.2375 0.2227 0.2299 3.56</cell></row><row><cell>kNN(275)</cell><cell cols="3">0.2102 0.1947 0.2022 kNN(275)</cell><cell cols="4">0.3413 0.3059 0.3226 3.40</cell></row><row><cell>kNN(421)</cell><cell cols="3">0.2749 0.0867 0.1318 kNN(421)</cell><cell cols="4">0.2787 0.1080 0.1557 2.26</cell></row><row><cell>kNN(rest)</cell><cell cols="3">0.1921 0.1627 0.1762 kNN(rest)</cell><cell cols="4">0.2007 0.1643 0.1807 61.41</cell></row><row><cell>FDT(1747)</cell><cell cols="3">0.5306 0.4169 0.4670 FDT(1747)</cell><cell cols="4">0.3592 0.2325 0.2823 19.90</cell></row><row><cell>FDT(2977)</cell><cell cols="3">0.4437 0.3622 0.3988 FDT(2977)</cell><cell cols="4">0.4162 0.3367 0.3722 9.48</cell></row><row><cell>FDT(483)</cell><cell cols="3">0.1684 0.2653 0.2060 FDT(483)</cell><cell cols="4">0.1642 0.2637 0.2024 3.56</cell></row><row><cell>FDT(275)</cell><cell cols="3">0.1887 0.1610 0.1738 FDT(275)</cell><cell cols="4">0.2531 0.1760 0.2076 3.40</cell></row><row><cell>FDT(421)</cell><cell cols="3">0.4044 0.1258 0.1920 FDT(421)</cell><cell cols="4">0.4328 0.1339 0.2045 2.26</cell></row><row><cell>FDT(rest)</cell><cell cols="3">0.2462 0.2133 0.2286 FDT(rest)</cell><cell cols="4">0.2502 0.2204 0.2344 61.41</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_102"><head>Table 6 .</head><label>6</label><figDesc>Best performance of various methods on bibtex training dataset. All values are averaged over 5 folds.</figDesc><table><row><cell>Method</cell><cell cols="2">Precision Recall F1-measure</cell></row><row><cell>kNN</cell><cell>0.3664 0.4307</cell><cell>0.3959</cell></row><row><cell cols="2">mpt+resource 0.3949 0.3765</cell><cell>0.3855</cell></row><row><cell>mpt+mix</cell><cell cols="2">0.4211 0.4014 0.4110</cell></row><row><cell>FolkRank</cell><cell>0.3222 0.4459</cell><cell>0.3741</cell></row><row><cell cols="3">DiffusionRank 0.3347 0.4630 0.3885</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_103"><head>Table 7 .</head><label>7</label><figDesc>Best performance of various methods on bookmark training dataset. All values are averaged over 5 folds.</figDesc><table><row><cell>Method</cell><cell cols="2">Precision Recall F1-measure</cell></row><row><cell>kNN</cell><cell>0.2855 0.2892</cell><cell>0.2873</cell></row><row><cell cols="2">mpt+resource 0.3345 0.2798</cell><cell>0.3047</cell></row><row><cell>mpt+mix</cell><cell>0.3606 0.3017</cell><cell>0.3285</cell></row><row><cell>FolkRank</cell><cell cols="2">0.3288 0.3309 0.3298</cell></row><row><cell cols="3">DiffusionRank 0.3772 0.3266 0.3501</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_104"><head></head><label></label><figDesc>. From the table, we find that the absolute values are much smaller than what are shown in Table 6 and 7.</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_105"><head>Table 8 .</head><label>8</label><figDesc>Evaluation result on test dataset of rsdc'09.</figDesc><table><row><cell cols="2">Tag Number Precision Recall F1-measure</cell></row><row><cell>1</cell><cell>0.1483 0.4229 0.2196</cell></row><row><cell>2</cell><cell>0.2301 0.3477 0.2769</cell></row><row><cell>3</cell><cell>0.2960 0.3113 0.3034</cell></row><row><cell>4</cell><cell>0.3418 0.2840 0.3102</cell></row><row><cell>5</cell><cell>0.3760 0.2601 0.3075</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_106"><head>Table 9 .</head><label>9</label><figDesc>Performance of FolkRank on bibtex training dataset. All values are averaged over 5 folds.</figDesc><table><row><cell>λ</cell><cell cols="2">max-it Precision Recall F1-measure</cell></row><row><cell>0.85</cell><cell>10 0.2943 0.4072</cell><cell>0.3417</cell></row><row><cell>0.5</cell><cell>10 0.3053 0.4225</cell><cell>0.3545</cell></row><row><cell>0.1</cell><cell>10 0.3198 0.4425</cell><cell>0.3713</cell></row><row><cell>0.01</cell><cell cols="2">10 0.3222 0.4459 0.3741</cell></row><row><cell>0.01</cell><cell cols="2">100 0.3222 0.4459 0.3741</cell></row><row><cell>0.001</cell><cell>10 0.3219 0.4455</cell><cell>0.3738</cell></row><row><cell>0.0001</cell><cell>10 0.3219 0.4455</cell><cell>0.3738</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_107"><head>Table 10 .</head><label>10</label><figDesc>Performance of FolkRank on bookmark training dataset. All values are averaged over 5 folds.</figDesc><table><row><cell>λ</cell><cell cols="2">max-it Precision Recall F1-measure</cell></row><row><cell>0.85</cell><cell>10 0.2989 0.3008</cell><cell>0.2998</cell></row><row><cell>0.5</cell><cell>10 0.3038 0.3058</cell><cell>0.3048</cell></row><row><cell>0.1</cell><cell>10 0.3198 0.3218</cell><cell>0.3208</cell></row><row><cell>0.01</cell><cell>10 0.3275 0.3297</cell><cell>0.3286</cell></row><row><cell>0.01</cell><cell>100 0.3275 0.3297</cell><cell>0.3286</cell></row><row><cell>0.001</cell><cell cols="2">10 0.3288 0.3309 0.3298</cell></row><row><cell>0.0001</cell><cell cols="2">10 0.3288 0.3309 0.3298</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_108"><head>Table 11 .</head><label>11</label><figDesc>Performance of DiffusionRank on bibtex training dataset. In all experiments, damping factor λ is set to λ = 0.85. All values are averaged over 5 folds.</figDesc><table><row><cell>γ</cell><cell cols="2">max-it Precision Recall F1-measure</cell></row><row><cell>2.0</cell><cell>10 0.3279 0.4537</cell><cell>0.3807</cell></row><row><cell>1.0</cell><cell>10 0.3331 0.4609</cell><cell>0.3867</cell></row><row><cell>0.1</cell><cell cols="2">10 0.3347 0.4630 0.3885</cell></row><row><cell>0.1</cell><cell cols="2">100 0.3347 0.4630 0.3885</cell></row><row><cell>0.01</cell><cell cols="2">10 0.3347 0.4630 0.3885</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_109"><head>Table 12 .</head><label>12</label><figDesc>Performance of DiffusionRank on bookmark training dataset. In all experiments, damping factor λ is set to λ = 0.85. All values are averaged over 5 folds.</figDesc><table><row><cell>γ</cell><cell cols="2">max-it Precision Recall F1-measure</cell></row><row><cell>2.0</cell><cell>10 0.3336 0.3357</cell><cell>0.3346</cell></row><row><cell>1.0</cell><cell>10 0.3370 0.3392</cell><cell>0.3381</cell></row><row><cell>0.1</cell><cell>10 0.3403 0.3425</cell><cell>0.3414</cell></row><row><cell>0.1</cell><cell>100 0.3403 0.3425</cell><cell>0.3414</cell></row><row><cell>0.01</cell><cell cols="2">10 0.3406 0.3428 0.3417</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_110"><head>Table 1 .</head><label>1</label><figDesc>Detail information about training dataset, provided by the contest</figDesc><table><row><cell>file</cell><cell># of lines information</cell></row><row><cell cols="2">bookmark 263,004 content id (matches tas.content id),</cell></row><row><cell></cell><cell>url hash (the URL as md5 hash),</cell></row><row><cell></cell><cell>url, description, extended description, date</cell></row><row><cell>bibtex</cell><cell>158,924 content id (matches tas.content id), journal, volume</cell></row><row><cell></cell><cell>chapter, edition, month, day, booktitle, editor, year</cell></row><row><cell></cell><cell>howPublished, institution, organization, publisher</cell></row><row><cell></cell><cell>address, school, series, bibtexKey, url, type, description</cell></row><row><cell></cell><cell>annote, note, pages, bKey, number, crossref, bibtexAbstract</cell></row><row><cell></cell><cell>simhash0, simhash1, simhash2, entrytype, title, author, misc</cell></row><row><cell>tas</cell><cell>1,401,104 userID, tag,</cell></row><row><cell></cell><cell>content id</cell></row><row><cell></cell><cell>(matches bookmark.content id or bibtex.content id)</cell></row><row><cell></cell><cell>content type (1 = bookmark, 2 = bibtex), date</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_111"><head>Table 2 .</head><label>2</label><figDesc>Performance of key word extraction method in every collection, while recommending at most 5 tags.Finding association rules in history recordsWe used three key factors in association rules, including support, confidence and interest. Every unique record is treated as the basket and the tags (X, Y , etc.) associated with every record are treated as the items in the basket. For every rule X → Y , support is the number of records that contain both X and Y . Confidence indicates the probability of</figDesc><table><row><cell>collection</cell><cell>BM25</cell><cell>TF-IDF</cell></row><row><cell></cell><cell cols="3">recall precision f-measure recall precision f-measure</cell></row><row><cell cols="2">bibtex original 0.0951 0.0561</cell><cell>0.0706 0.0989 0.0592</cell><cell>0.0741</cell></row><row><cell cols="2">bibtex parsed 0.1663 0.1059</cell><cell>0.1294 0.1800 0.1158</cell><cell>0.1409</cell></row><row><cell cols="2">bookmark more 0.1186 0.0940</cell><cell>0.1049 0.1189 0.0943</cell><cell>0.1052</cell></row><row><cell cols="4">alternative approaches to deeply analyze association rules, which are found in</cell></row><row><cell cols="4">history information. It does help to extract tags which are more likely to be used</cell></row><row><cell>by users.</cell><cell></cell><cell></cell></row></table><note>Y in this record if X already associates with the record, i.e., P (Y |X). Interest is P (Y |X) − P (Y ), showing how much more possible that X and Y associating with the record together.</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_112"><head>Table 3 .</head><label>3</label><figDesc>Sample Association rules found in training dataset</figDesc><table><row><cell cols="2">bookmark</cell><cell></cell><cell></cell><cell cols="2">bibtex</cell><cell></cell><cell></cell></row><row><cell>X → Y</cell><cell cols="3">confidence support interest</cell><cell>X → Y</cell><cell cols="3">confidence support interest</cell></row><row><cell>blog → sof tware</cell><cell>0.0541</cell><cell>291</cell><cell>0.0454</cell><cell>systems → algorithms</cell><cell>0.2886</cell><cell>295</cell><cell>0.2757</cell></row><row><cell>blogs → blogging</cell><cell>0.1345</cell><cell>291</cell><cell>0.1333</cell><cell>algorithms → systems</cell><cell>0.0492</cell><cell>295</cell><cell>0.0470</cell></row><row><cell>blogging → blogs</cell><cell>0.2910</cell><cell>291</cell><cell>0.2885</cell><cell>systems → genetic</cell><cell>0.2847</cell><cell>291</cell><cell>0.2721</cell></row><row><cell>artery → cardiology</cell><cell>0.9510</cell><cell>291</cell><cell>0.9506</cell><cell>genetic → systems</cell><cell>0.0497</cell><cell>291</cell><cell>0.0475</cell></row><row><cell>photos → photography</cell><cell>0.3149</cell><cell>290</cell><cell cols="2">0.3138 tagging → f olksonomy</cell><cell>0.5097</cell><cell>288</cell><cell>0.5085</cell></row><row><cell>photography → photos</cell><cell>0.3142</cell><cell>290</cell><cell cols="2">0.3131 f olksonomy → tagging</cell><cell>0.5115</cell><cell>288</cell><cell>0.5103</cell></row><row><cell>learning → f oodcooking</cell><cell>0.1004</cell><cell>290</cell><cell>0.1000</cell><cell>genetic → and</cell><cell>0.0466</cell><cell>273</cell><cell>0.0441</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_113"><head></head><label></label><figDesc>According to the above equations, the algorithm to calculate P a (Y ), which is called Assoc(Y ) in this paper, is shown in Algorithm 2.</figDesc><table><row><cell>end for</cell></row><row><cell>end for</cell></row><row><cell>end for</cell></row></table><note>Algorithm 2 To calculate P a (X), by using association rules for all documents in the collection do for all term X in the document do for all association rule X → Y do Pa(Y )+ = (conf idence of X → Y ) * P k (X); {//P k (X) is calculated by Algorithm 1}</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_114"><head>Table 4 .</head><label>4</label><figDesc>Performance for only using TF-IDF results, linearly combining results of TF-IDF &amp; association rules, and common &amp; combine the two results, for the top N tag recommendations</figDesc><table><row><cell></cell><cell>TF-IDF</cell><cell></cell><cell></cell><cell cols="2">linearly combining results</cell><cell></cell><cell cols="2">common &amp; combine</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell>bibtex original</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="9">Top N recall precision f-measure Top N recall precision f-measure Top N recall precision f-measure</cell></row><row><cell>1</cell><cell>0.0199 0.0636</cell><cell>0.0304</cell><cell>1</cell><cell>0.0339 0.1105</cell><cell>0.0519</cell><cell>1</cell><cell>0.0344 0.1123</cell><cell>0.0527</cell></row><row><cell>2</cell><cell>0.0378 0.0610</cell><cell>0.0467</cell><cell>2</cell><cell>0.0593 0.0979</cell><cell>0.0739</cell><cell>2</cell><cell>0.0619 0.1018</cell><cell>0.0770</cell></row><row><cell>3</cell><cell>0.0579 0.0603</cell><cell>0.0591</cell><cell>3</cell><cell>0.0824 0.0900</cell><cell>0.0860</cell><cell>3</cell><cell>0.0848 0.0927</cell><cell>0.0886</cell></row><row><cell>4</cell><cell>0.0787 0.0598</cell><cell>0.0680</cell><cell>4</cell><cell>0.1046 0.0849</cell><cell>0.0937</cell><cell>4</cell><cell>0.1065 0.0867</cell><cell>0.0956</cell></row><row><cell>5</cell><cell>0.0989 0.0592</cell><cell>0.0741</cell><cell>5</cell><cell>0.1244 0.0802</cell><cell>0.0975</cell><cell>5</cell><cell>0.1264 0.0816</cell><cell>0.0992</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell>bibtex parsed</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="9">Top N recall precision f-measure Top N recall precision f-measure Top N recall precision f-measure</cell></row><row><cell>1</cell><cell>0.0708 0.2033</cell><cell>0.1050</cell><cell>1</cell><cell>0.0728 0.2106</cell><cell>0.1081</cell><cell>1</cell><cell>0.0723 0.2155</cell><cell>0.1083</cell></row><row><cell>2</cell><cell>0.1138 0.1710</cell><cell>0.1367</cell><cell>2</cell><cell>0.1171 0.1802</cell><cell>0.1419</cell><cell>2</cell><cell>0.1212 0.1871</cell><cell>0.1471</cell></row><row><cell>3</cell><cell>0.1438 0.1487</cell><cell>0.1462</cell><cell>3</cell><cell>0.1527 0.1605</cell><cell>0.1565</cell><cell>3</cell><cell>0.1549 0.1635</cell><cell>0.1591</cell></row><row><cell>4</cell><cell>0.1665 0.1316</cell><cell>0.1470</cell><cell>4</cell><cell>0.1771 0.1425</cell><cell>0.1580</cell><cell>4</cell><cell>0.1778 0.1432</cell><cell>0.1586</cell></row><row><cell>5</cell><cell>0.1800 0.1158</cell><cell>0.1409</cell><cell>5</cell><cell>0.1959 0.1281</cell><cell>0.1549</cell><cell>5</cell><cell>0.1968 0.1291</cell><cell>0.1559</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell>bookmark more</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="9">Top N recall precision f-measure Top N recall precision f-measure Top N recall precision f-measure</cell></row><row><cell>1</cell><cell>0.0388 0.1285</cell><cell>0.0596</cell><cell>1</cell><cell>0.0449 0.1547</cell><cell>0.0696</cell><cell>1</cell><cell>0.0460 0.1599</cell><cell>0.0715</cell></row><row><cell>2</cell><cell>0.0693 0.1172</cell><cell>0.0871</cell><cell>2</cell><cell>0.0846 0.1400</cell><cell>0.1055</cell><cell>2</cell><cell>0.0872 0.1487</cell><cell>0.1099</cell></row><row><cell>3</cell><cell>0.0919 0.1080</cell><cell>0.0993</cell><cell>3</cell><cell>0.1133 0.1286</cell><cell>0.1205</cell><cell>3</cell><cell>0.1165 0.1358</cell><cell>0.1254</cell></row><row><cell>4</cell><cell>0.1077 0.1001</cell><cell>0.1038</cell><cell>4</cell><cell>0.1375 0.1202</cell><cell>0.1283</cell><cell>4</cell><cell>0.1415 0.1255</cell><cell>0.1330</cell></row><row><cell>5</cell><cell>0.1189 0.0943</cell><cell>0.1052</cell><cell>5</cell><cell>0.1581 0.1132</cell><cell>0.1319</cell><cell>5</cell><cell>0.1623 0.1172</cell><cell>0.1361</cell></row></table><note>high P c (X) for recommendation. The total number of tags to recommend is controlled by k, the number of tags to check in the common step is common-no, and the number of tags to extract in the combine step is combine-no. Detailed steps are shown in Algorithm 4.</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_115"><head>Table 5 .</head><label>5</label><figDesc>Priority to combine resultsTags from records that has exact match with same user and same bookmark/bibtex Tags from records that has match with same user Tags from records that has match with same resource (bookmark url or bibtex publication)common &amp; combine results of bibtex parsed common &amp; combine results of bibtex original common &amp; combine results of bookmark more</figDesc><table><row><cell>Priority</cell><cell>Method</cell></row><row><cell>Higher to lower</cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_116"><head>Table 6 .</head><label>6</label><figDesc>Performance for without checking history records, resource match with higher priority and user match with higher priority, for the top N tag recommendations</figDesc><table><row><cell cols="3">without checking history records</cell><cell></cell><cell cols="2">resource match higher</cell><cell></cell><cell cols="2">user match higher</cell></row><row><cell cols="9">Top N recall precision f-measure Top N recall precision f-measure Top N recall precision f-measure</cell></row><row><cell>1</cell><cell>0.0415 0.1395</cell><cell>0.0639</cell><cell>1</cell><cell>0.0835 0.2312</cell><cell>0.1226</cell><cell>1</cell><cell>0.0867 0.2396</cell><cell>0.1273</cell></row><row><cell>2</cell><cell>0.0783 0.1305</cell><cell>0.0979</cell><cell>2</cell><cell>0.1344 0.2143</cell><cell>0.1652</cell><cell>2</cell><cell>0.1374 0.2220</cell><cell>0.1698</cell></row><row><cell>3</cell><cell>0.1059 0.1204</cell><cell>0.1126</cell><cell>3</cell><cell>0.1667 0.1980</cell><cell>0.1810</cell><cell>3</cell><cell>0.1684 0.2064</cell><cell>0.1855</cell></row><row><cell>4</cell><cell>0.1292 0.1115</cell><cell>0.1197</cell><cell>4</cell><cell>0.1915 0.1866</cell><cell>0.1890</cell><cell>4</cell><cell>0.1916 0.1954</cell><cell>0.1935</cell></row><row><cell>5</cell><cell>0.1510 0.1046</cell><cell>0.1235</cell><cell>5</cell><cell>0.2118 0.1778</cell><cell>0.1933</cell><cell>5</cell><cell>0.2104 0.1871</cell><cell>0.1981</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_117"><head>Table 7 .</head><label>7</label><figDesc>Best performance on training dataset, only for bookmarks and only for publications, for the top N tag recommendations</figDesc><table><row><cell></cell><cell>all resources</cell><cell></cell><cell></cell><cell cols="2">only for bookmark</cell><cell></cell><cell>only for bibtex</cell><cell></cell></row><row><cell cols="9">Top N recall precision f-measure Top N recall precision f-measure Top N recall precision f-measure</cell></row><row><cell>1</cell><cell>0.0867 0.2396</cell><cell>0.1273</cell><cell>1</cell><cell>0.0771 0.2364</cell><cell>0.1163</cell><cell>1</cell><cell>0.1025 0.2448</cell><cell>0.1445</cell></row><row><cell>2</cell><cell>0.1374 0.2220</cell><cell>0.1698</cell><cell>2</cell><cell>0.1296 0.2215</cell><cell>0.1635</cell><cell>2</cell><cell>0.1504 0.2228</cell><cell>0.1796</cell></row><row><cell>3</cell><cell>0.1684 0.2064</cell><cell>0.1855</cell><cell>3</cell><cell>0.1613 0.2056</cell><cell>0.1808</cell><cell>3</cell><cell>0.1803 0.2076</cell><cell>0.1930</cell></row><row><cell>4</cell><cell>0.1916 0.1954</cell><cell>0.1935</cell><cell>4</cell><cell>0.1843 0.1932</cell><cell>0.1887</cell><cell>4</cell><cell>0.2038 0.1990</cell><cell>0.2014</cell></row><row><cell>5</cell><cell>0.2104 0.1871</cell><cell>0.1981</cell><cell>5</cell><cell>0.2035 0.1841</cell><cell>0.1933</cell><cell>5</cell><cell>0.2218 0.1921</cell><cell>0.2059</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_118"><head>Table 1 .</head><label>1</label><figDesc>Different node set sizes of the Bibsonomy dataset for p-core levels 1 and 2.</figDesc><table><row><cell>p-core</cell><cell>|E|</cell><cell cols="2">|BM | |BMBIB| |BMURL| |I|</cell><cell>|T |</cell><cell>|U |</cell></row><row><cell>1</cell><cell cols="5">1,401,104 421,928 263,004 158,924 378,378 93,756 3,617</cell></row><row><cell>2</cell><cell cols="2">253,615 64,120 41,268</cell><cell cols="3">22,852 22,389 13,276 1,185</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_119"><head>Table 1 :</head><label>1</label><figDesc>Examples of tag sources</figDesc><table><row><cell></cell><cell>TITLE</cell><cell>TAGS</cell><cell>SOURCE</cell></row><row><cell></cell><cell></cell><cell></cell><cell>OF TAGS</cell></row><row><cell>37</cell><cell>SVG: Adobe</cell><cell>adobe, svg</cell><cell>title</cell></row><row><cell>787</cell><cell>SourceForge.net: delicious-</cell><cell>api java, delicious</cell><cell>title</cell></row><row><cell></cell><cell>java</cell><cell></cell><cell></cell></row><row><cell>173</cell><cell>Reassessing Working</cell><cell>psycholinguistics,</cell><cell>concept or</cell></row><row><cell></cell><cell>Memory: Comment on Just</cell><cell>review,</cell><cell>topic</cell></row><row><cell></cell><cell>and Carpenter (1992) and</cell><cell>workingmemory</cell><cell></cell></row><row><cell></cell><cell>Waters and Caplan (1996)</cell><cell></cell><cell></cell></row><row><cell>293</cell><cell>A Semantic Web Primer</cell><cell>swss0603,</cell><cell>specific to</cell></row><row><cell></cell><cell></cell><cell>ontolex2006,</cell><cell>the user</cell></row><row><cell></cell><cell></cell><cell>semwebss06, swss0602</cell><cell></cell></row><row><cell>293</cell><cell>The ABCDE Format Enabling</cell><cell>semwiki2006,swikig,w</cell><cell>specific to</cell></row><row><cell></cell><cell>Semantic Conference</cell><cell>iki, eswc2006,</cell><cell>the user</cell></row><row><cell></cell><cell>Proceedings</cell><cell>semantic</cell><cell></cell></row><row><cell>293</cell><cell>Learning of Ontologies for the</cell><cell>ontologylearning,</cell><cell>specific to</cell></row><row><cell></cell><cell>Web: the Analysis of Existent</cell><cell>semanticweb,semwebs</cell><cell>the user</cell></row><row><cell></cell><cell>Approaches</cell><cell>s06, sw0809,</cell><cell></cell></row><row><cell></cell><cell></cell><cell>sw080912, swss0609</cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_120"><head>Table 2 :</head><label>2</label><figDesc>Notations. recommendation system. The candidate set C is composed with two subset, C 1 and C 2 , i.e. C = C 1 ∪ C 2 .</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_121"><head>Table 4 :</head><label>4</label><figDesc>performance of ACT model on the test data, the numbers are shown in the following format: recall/precision/f-measure4.3.3 Performance after combinationIn table 5, we show the performance after the combination of these two models. From the table, we can see that after combination, our recommendation system works a little better than the Language model and has a highest f-measure of 14.398% when recommending five tags.</figDesc><table><row><cell></cell><cell>bookmark(%)</cell><cell>bibtex(%)</cell><cell>overall(%)</cell></row><row><cell>1</cell><cell>2.142/6.800/3.258</cell><cell>0.944/2.758/1.406</cell><cell>1.415/4.346/2.135</cell></row><row><cell>2</cell><cell>3.523/5.628/4.333</cell><cell>1.431/2.320/1.770</cell><cell>2.253/3.620/2.778</cell></row><row><cell>3</cell><cell>4.519/4.825/4.667</cell><cell>1.885/2.076/1.976</cell><cell>2.920/3.156/3.034</cell></row><row><cell>4</cell><cell>5.179/4.196/4.636</cell><cell>2.167/1.868/2.007</cell><cell>3.351/2.783/3.041</cell></row><row><cell>5</cell><cell>5.829/3.815/4.612</cell><cell>2.466/1.721/2.027</cell><cell>3.788/2.544/3.044</cell></row><row><cell>6</cell><cell>6.418/3.536/4.560</cell><cell>2.724/1.582/2.002</cell><cell>4.175/2.350/3.077</cell></row><row><cell>7</cell><cell>6.870/3.257/4.419</cell><cell>2.977/1.483/1.980</cell><cell>4.507/2.180/2.939</cell></row><row><cell>8</cell><cell>7.377/3.059/4.324</cell><cell>3.205/1.393/1.942</cell><cell>4.844/2.048/2.878</cell></row><row><cell>9</cell><cell>7.849/2.891/4.225</cell><cell>3.557/1.346/1.953</cell><cell>5.244/1.953/2.846</cell></row><row><cell>10</cell><cell>8.289/2.746/4.126</cell><cell>3.721/1.276/1.900</cell><cell>5.516/1.854/2.775</cell></row><row><cell></cell><cell></cell><cell>final result(%)</cell><cell></cell></row><row><cell></cell><cell>1</cell><cell>4.624/15.271/7.099</cell><cell></cell></row><row><cell></cell><cell>2</cell><cell>7.753/14.550/10.116</cell><cell></cell></row><row><cell></cell><cell>3</cell><cell>10.626/14.900/12.405</cell><cell></cell></row><row><cell></cell><cell>4</cell><cell>12.738/14.944/13.753</cell><cell></cell></row><row><cell></cell><cell>5</cell><cell>13.916/14.915/14.398</cell><cell></cell></row><row><cell>Table</cell><cell></cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_122"><head>Table 3 :</head><label>3</label><figDesc>the general statistical information about the test dataset</figDesc><table><row><cell>w 0</cell></row><row><cell>end</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">http://www.kde.cs.uni-kassel.de/ws/dc09/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">http://www.bibsonomy.org</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">http://www.nokia.com/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_3">http://www.tagora-project.eu/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_4">Since the results of PWA* and WA* are very similar, we just report on WA*.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_5">Delicious -Social bookmarking, http://delicious.com/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_6">CiteULike -Scholarly reference management and discovery, http://www.citeulike.com/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_7">Flickr -Photo sharing, http://www.flickr.com/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_8">YouTube -Video sharing, http://www.youtube.com/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_9">Last.fm -Personal online radio, http://www.last.fm/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_10">ECML PKDD 2009 Discovery Challenge, http://www.kde.cs.uni-kassel.de/ws/dc09/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_11">BibSonomy -Social bookmark and publication sharing, http://www.bibsonomy.org/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_12">Apache Lucene -Open-source Information Retrieval library, http://lucene.apache.org/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_13">Relative frequency of a tag in a random collection of 603 750 downloaded from the Delicious.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_14">Rank of a corresponding tag in the BibSonomy.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_15">http://www.linuxinsider.com/perl/syndication/rssfull.pl</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_16"> delicious.com   </note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_17">www.flickr.com</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_18">www.last.fm</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_19">Informational Channels of FolksonomiesThe model of a folksonomy suggests several informational channels which may be exploited by data mining applications such as tag recommenders. The relation between users, resources and tags generate a complex network of interrelated items as shown in Figure1.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_20">www.bibsonomy.org</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_21">http://www.bibsonomy.org</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_22">More details about tasks can be found on Challenge's site at http://www.kde.cs.uni-kassel.de/ws/dc09/#tasks</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_23">More details about datasets can be found at http://www.kde.cs.uni-kassel.de/ws/dc09/dataset</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_24">http://www.kde.cs.uni-kassel.de/ws/dc09</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_25">http://fp7.okkam.org/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_26">http://livingknowledge-project.eu/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_27">http://www.bibsonomy.org/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_28">formerly del.icio.us, http://delicious.com/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_29"> [1]  also exploited these kinds of information sources for tag recommendation. We extend this approach by extracting keywords from not only resource title but also other resource descriptions.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_30">It is because average importance values of keywords are different according to extracted columns.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_31">MC(k, d) is equal to EC(k, d) if d is tagged with k, 0 otherwise.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_32">For the remainder of this paper, we will refer to this process as "(automatic) tagging"</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_33">http://www.bibsonomy.org</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_34">http://incubator.apache.org/pdfbox/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_35">http://en.wikipedia.org</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_36">projection π and selection σ operate on multisets without removing duplicate tuples</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_37">Unless stated explicitly otherwise, we recommend at least one tag and at most the number of tags annotated to a resource</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_38">Our submission to the DC09 challenge was based on 2500 latent topics without combination with most frequent tags, which achieved an F-measure of 0.098.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_39">Our submission to the DC09 challenge was based on</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5000" xml:id="foot_40">topics without combination with the most frequent tags and no limit on the number of recommended tags. This achieved an F-measure of 0.258.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_41">http://del.icio.us</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_42">http://www.flickr.com/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_43">http://www.bibsonomy.org/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_44">http://www.kde.cs.uni-kassel.de/ws/dc09</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_45">http://wordnet.princeton.edu</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_46">http://www.kde.cs.uni-kassel.de/ws/dc09</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_47">http://www.bibsonomy.org/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="10" xml:id="foot_48">http://www.lextek.com/manuals/onix/stopwords1.html</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="11" xml:id="foot_49">http://www.kde.cs.uni-kassel.de/ws/dc09</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_50">http://bibsonomy.org/help/about/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_51">http://del.icio.us/about/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_52">http://flickr.com/about/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_53">http://technorati.com/about/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_54">http://www.kde.cs.uni-kassel.de/ws/dc09/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_55">http://www.kde.cs.uni-kassel.de/ws/dc09/dataset</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_56">http://www.citeulike.org/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_57">http://www.kde.cs.uni-kassel.de/ws/rsdc08/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_58">available at http://www.csie.ntu.edu.tw/∼cjlin/liblinear/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_59">http://www.kde.cs.uni-kassel.de/ws/dc09/dataset</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_60">http://www.bibsonomy.org</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_61">in the sense that the tags indeed help users to find or organize resources, for recent results on tag quality compare[2, 3].</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_62">Group 1: Mrosek, Bussmann, Albers, Posdziech.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_63">Group 2: Hengefeld, Opperman, Robert, Spira.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_64">Not a very successful idea -first evaluations show that this was rather counterproductive.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_65">value = value * Math.pow((1.0 + 3*(count -1)/10), count);</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_66">The following section documents the work of group 1 only</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_67">and the anonymous reviewers and Wolfram Conen for their valuable comments.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_68">A Hybrid Tag RecommenderIn our experiment, we use a hybrid recommender as described in detail in Algorithm 4. The recommender checks if a given resource exists in the training data.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_69">http://www.kde.cs.uni-kassel.de/ws/dc09</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_70">http://www.bibsonomy.org</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_71">http://lucene.apache.org/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_72">http://www.cs.waikato.ac.nz/ml/weka/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_73">http://www.flickr.com</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_74">http://www.youtube.com</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_75">http://delicious.com/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_76">http://www.last.fm/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_77">http://www.bibsonomy.org/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_78">http://www.kde.cs.uni-kassel.de/ws/dc09</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_79">http://lucene.apache.org</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_80">http://nlp.uned.es/ jperezi/Lucene-BM25/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" xml:id="foot_81">The provided training data contains three files: bibtex, bookmark and tas. The bibtex and bookmark files describe the content of the links and BibTeX entries, respectively. The tas file contains the tag assignments. Also provided was the post-core at level 2[3], which contained a reduced set, which contained only</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_82">http://lucene.apache.org</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_83">Conclusions and Future WorkIn this paper, we proposed a tag recommendation system using keywords in the page content and association rules from history records. If the record resource</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_84">http://delicious.com</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_85">http://bibsonomy.org</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_86">http://citeulike.org</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_87">http://flickr.com</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_88">http://www.youtube.com</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_89">http://www.kde.cs.uni-kassel.de/ws/dc09</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_90">The p-core of a folksonomy graph has the characteristic that all contained nodes appear in at least p bookmarks. See[4]  for details.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="10" xml:id="foot_91">Note that a small percentage of user item combinations found in the given dataset occur in more than one bookmark.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_92">http://www.bibsonomy.org</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_93">http://www.flick.com</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_94">http://del.icio.us</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_95">http://www.amazon.com</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_96">http://www.last.fm</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_97">http://www.eBay.com</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7">Acknowledgements</head><p>This work is supported by CNPq, an institution of Brazilian Government for scientific and technologic development, and the X-Media project (www.x-media-project.org) sponsored by the European Commission as part of the Information Society Technologies (IST) programme under EC grant number IST-FP6-026978. The authors also gratefully acknowledge the partial co-funding of their work through the European Commission FP7 project MyMedia (www.mymediaproject.org) under the grant agreement no. 215006. For your inquiries please contact info@mymediaproject.org.</p><p>Acknowledgments. This research was supported by the European Commission under contracts FP6-027122-SALERO, FP6-033715-MIAUCE and FP6-045032 SEMEDIA. The expressed content is the view of the authors but not necessarily the view of SALERO, MIAUCE and SEMEDIA projects as a whole.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgement</head><p>Thanks to Zhen Liao for his helpful discussions and suggestions for this paper. This paper is supported by the National Natural Science Foundation of China under the grant 60673009 and China National Hanban under the grant 2007-433. 8 Acknowledgments This work was supported in part by the National Science Foundation Cyber Trust program under Grant IIS-0430303 and a grant from the Department of Education, Graduate Assistance in the Area of National Need, P200A070536. Acknowledgments This paper is part of the 03ED316/8.3.1. research project, implemented within the framework of the "Reinforcement Programme of Human Research Manpower" (PENED) and co-financed by National and Community Funds (20% from the Greek Ministry of Development-General Secretariat of Research and Technology and 80% from E.U.-European Social Fund). Acknowledgments. This work is partially supported by the EU Large-scale Integrating Projects OKKAM 2 -Enabling a Web of Entities (contract no. ICT-215032), and LivingKnowledge 3 (contract no. 231126) Acknowledgements This work was supported in part by the Seoul Development Institute through Seoul R&amp;BD Program (GS070167C093112) and in part by the Ministry of Culture, Sports and Tourism of Korea through CT R&amp;D Program (20912050011098503004).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Acknowledgments</head><p>This work was supported in part by the EU project IST 45035 -Platform for searcH of Audiovisual Resources across Online Spaces (PHAROS).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgement</head><p>This paper is supported by the National Natural Science Foundation of China under the grant 60673009 and China National Hanban under the grant 2007-433. The authors thank Chin-Yew Lin at Microsoft Research Asia for his valuable comments to this paper. Thanks also to Jie Liu, Yang Wang and Min Lu for their helpful discussions and suggestions.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7">Acknowledgments</head><p>The authors would like to thank Nicolas Neubauer for useful discussions, comments and suggestions in writing this paper. The first author was funded partly by a scholarship by the DAAD.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7">Acknowledgements</head><p>The author acknowledges Heikki Kallasjoki's technical assistance and Mari-Sanna Paukkeri's comments. This work was supported by the Academy of Finland through the Adaptive Informatics Research Centre that is a part of the Finnish Centre of Excellence Programme.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="8">Acknowledgements</head><p>The authors gratefully acknowledge the partial co-funding of their work through the European Commission FP7 project MyMedia (www.mymediaproject.org) under the grant agreement no. 215006. For your inquiries please contact info@mymediaproject.org.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This work is supported by the National Science Foundation of China under Grant No. 60621062, 60873174 and the National 863 High-Tech Project under Grant No. 2007AA01Z148.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This work was supported in part by a grant from the National Science Foundation under award IIS-0545875.</p></div>
			</div>

			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A Two-Level Learning Hierarchy of Concept</head><p>Based Keyword Extraction for Tag Recommendations</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Hendri Murfi and Klaus Obermayer</head><p>Neural Information Processing Group, TU Berlin Franklinstr. 28/29, 10587 Berlin, Germany {henri,oby}@cs.tu-berlin.de http://ni.cs.tu-berlin.de</p><p>Abstract. Textual contents associated to resources are considered as sources of candidate tags to improve the performance of tag recommenders in social tagging systems. In this paper, we propose a twolevel learning hierarchy of a concept based keyword extraction method to filter the candidate tags and rank them based on their occurrences in concepts existing in the given resources. Incorporating user-created tags to extract the hidden concept-document relationships distinguishes the two-level from the one-level learning version, which extracts concepts directly using terms existing in textual contents. Our experiment shows that a multi-concept approach, which considers more than one concept for each resource, improves the performance of a single-concept approach, which takes into account just the most relevant concept. Moreover, the experiments also prove that the proposed two-level learning hierarchy gives better performances than one of the one-level version.</p><p>Projections π U R Y ∈ 0, 1 |U |×|R| , (π U R Y ) u,r := 1 iff ∃t ∈ T s.t. (u, r, t) ∈ Y and π U T Y ∈ 0, 1 |U |×|T | , (π U T Y ) u,t := 1 iff ∃r ∈ R s.t. (u, r, t) ∈ Y let us define the "tag neighbourhood" and "resource neighbourhood" of the users. The set of k nearest neighbours for a user u using the neighbourhood matrix X is</p><p>where sim is the cosine similarity sim(x, y)</p><p>The set of recommendations for a given user-resource pair (u, r) is</p><p>where δ(v, r, t) := iff(v, r, t) ∈ Y .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Baseline Methods</head><p>The following are a collection of simple recommendation methods, which do not produce very good recommendations and have few redeeming qualities except that they are computationally inexpensive.</p><p>Popular tags for a resource. If the users of the folksonomy are homogenous, this method can be expected to perform almost as well as CF methods. However, if the users have very different tagging habits or if people use different tags from different languages, performance for the minorities can be expected to suffer.</p><p>Popular tags for a user. Some users use relatively few but obscure tags, which means that the popular tags for resource -recommender will not work. Collaborative recommendations also will not work well, as the user will probably have very few applicable "tag neighbours" and the "resource neighbours" will most likely not use the same tags. For example, user 483 used the tag "allgemein" a total of 2237 times in the 9003 posts. In other words, given a post by this user at random, there is almost a 25% chance it is tagged "allgemein".</p><p>Globally popular tags. Recommending the most used tags is perhaps the simplest possible method.</p><p>We used several variants of the aforementioned recommenders. These and the method used to combine the recommendations are described in chapter 4.1.</p><p>After the folding we apply cosine similarity to compare two tag vectors:</p><p>We tried different weighted ensembles of the baseline models using the value estimate ensembling method. Even though these ensembles produce quite good results, in our experiments they did not outperform the factor models and furthermore adding baselines to the factor models did not result in a significant improvement of the factor models. Thus our final submission only consists of the factor models.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Adaptive List Length</head><p>In contrast to the usual evaluation scheme of tag recommendation, in this challenge the recommender was free to choose the length of the list of the recommendations in a range from a length of 0 to 5. The evaluation functions are: Where # u,i is the number of tags the recommender estimates for a post. There are three simple ways to estimate # u,i :</p><p>-Global estimate:</p><p>-User estimate: Based on the graph, we can employ various graph-based ranking methods to recommend tags. In this paper, we first introduce two existing methods, including "most popular tags" and "FolkRank". Furthermore, we propose to use a new ranking model, DiffusionRank, for graph-based tag suggestion.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Most Popular Tags</head><p>We first introduce a simple but effective method for tag suggestion. Some notations are given as below, which is identical with <ref type="bibr">[2]</ref>. For a user u ∈ U , we denote all his/her tag assignments as Y u := Y ({u} × T × R). Accordingly, we have Y r and Y t . Based on the same principle, we can define Y u,t := Y ({u} × {t} × R) for u ∈ U and t ∈ T . We also have Y t,r accordingly. Furthermore, we denote all tags that user u ∈ U have assigned as</p><p>There are variants of "most popular tags" as shown in <ref type="bibr">[8]</ref>, which are usually restricted in different statistical range. For example, most popular tags of folksonomy recommends the most popular tags of the whole set of folksonomy. Therefore, it recommends the same set of tags for any user and resource, which suffers from cold-start problems and has no consideration on personalization.</p><p>A reasonable variant of "most popular tags" is recommending the tags that globally are most specific to the resource. The method is named as most popular tags by resource:</p><p>Since users might have specific preferences for some tags, which should have been used by him/her, thus we can use the most popular tags by user. As shown in <ref type="bibr">[8]</ref>, the performance is poor if we use most popular tags by user in isolation. If we mix the most popular tags of user and resource, the performance will be much better than each of them. The simplest way to mix the effect of users and resources on tags is to add the counts and then sort:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">FolkRank</head><p>FolkRank is originally proposed in <ref type="bibr">[2]</ref> which is based on user-resource-tag tripartite graph. In FolkRank, two random surfer model is employed on the tripartite graph. The ranking values of vertices are computed using the following formula:</p><p>where P R(v i ) is the PageRank value and p vi is the preference to v i . Suppose we have an adjacent matrix A to represent the graph G F : For each group of test data, we have 4 different models, they are kNN trained by all users, kNN trained by this group, FDT trained by all users and FDT trained by this user. As the result shows, for groups of super users, kNN-based models have best performance. For common users (the rest group), FDT-based models performs better. This result follows our intuition. In this data set, super users have different tag preference than common users. kNN suggest tags using most similar resources, it is less affected by the overall distribution of resources, so it fits to the . FDT relies on the global statistics of feature-tag relationship, it is less effective to fit a special user's preference. In practical situation, we can get the best performance by choosing different model for different group of users.</p><p>One interesting observation is about the user #2732. When trained with all posts, FDT performs much better(0.6308 vs 0.2300) on #2732 than trained with #2732's own posts. We examined the posts of #2732, found that many posts contains only three tags: genetic, programming and algorithm, and the number of posts by #2732 is large. When we use all posts to train FDT(TFITF+MI), these three tags have a large Mutual Information value with many features, especially the user id feature "UID-2732", so FDT can predict tags for posts of #2732 with high accuracy. When trained only with #2732's posts, the Mutual Information between features and these three tags is much smaller, since these three tags appears everywhere and can be seen as stopwords in tags. Small Mutual Information of these three tags means FDT will make wrong prediction about most posts of #2732, which leads to a decreasing in F1-measure.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Department of Computer Science and Engineering</head><p>Lehigh University, Bethlehem, PA 18015 USA {jiw307, lih307, davison}@cse.lehigh.edu</p><p>Abstract. While a webpage usually contains hundreds of words, there are only two to three tags that would typically be assigned to this page. Most tags could be found in related aspects of the page, such as the page own content, the anchor texts around the page, and the user's own opinion about the page. Thus it is not an easy job to extract the most appropriate two to three tags to recommend for a target user.</p><p>In addition, the recommendations should be unique for every user, since everyone's perspective for the page is different. In this paper, we treat the task of recommending tags as to find the most likely tags that would be chosen by the user. We first applied the TF-IDF algorithm on the limited description of the page content, in order to extract the keywords for the page. Based on these top keywords, association rules from history records are utilized to find the most probable tags to recommend. In addition, if the page has been tagged before by other users or the user has tagged other resources before, that history information is also exploited to find the most appropriate recommendations.</p><p>Algorithm 3 To calculate P c (X), by linearly combining results from TF-IDF &amp; association rules for all documents in the collection do T F − IDF max = maximum T F − IDF (X) for all terms in this document Assocmax = maximum Assoc(X) for all terms in this document for all term X in the document do    Node degree distributions within the full Bibsonomy dataset (a) and the dataset at p-core level 2. Node degree distributions found within folksonomies generally exhibit power law characteristics with few highly connected nodes and many nodes with low occurrence. This characteristic is obtained when cutting the graph to its level 2-core.</p><p>reduces the overall accuracy of our recommender. However, we believe this loss to be compensated by the general applicability of our approach and the improved validity of the presented results.</p><p>Figure <ref type="figure">1</ref> shows the node degree distributions of the full dataset and its 2core. As previously reported for other folksonomies (e.g. <ref type="bibr">[2]</ref>[7]), we find the node degree distributions of Bibsonomy to exhibit power law characteristics, with very few and very frequent nodes on one end and many infrequent nodes on the other. This characteristic is basically obtained when reducing G to its 2-core. However, we find that this reduction drastically reduces the impact of some previously very influential users. These users tend to have a rather "organizational" background and bookmark large collections of similar resources which are of no interest to other users. Most of these resources therefore fall victim to the p-core reduction, which also explains the drastic cut on items in Table <ref type="table">1</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Tag vocabularies</head><p>As users bookmark items only once 10 , we cannot estimate P (t|i, u) directly from previous observations. Instead, we base our estimates on other distributions. The most basic sources are the overall tag distribution and the tags previously assigned by the user, or to the item.</p><p>Global tag distributions. This is the distribution of all observed tag assignments within the training data. By definition, each tag of the test data has neighbor of the resource r, noted as neigh(r). C 2 is compose of the tags previous posted to the neigh(r * ). For each t 2 ∈ C 2 , we have the following generative probability:</p><p>(2)</p><p>where n(t 2 ,r) is the number of times that tag t 2 has been posted to the resource r and n(t 2 ,R) is the number of times that tag t 2 has been posted in the training data. T r is the number of tags posted to the resource r and λ is the Dirichlet smoothing factor and is commonly set according to the average document length, i.e. T /|R| Now, we have the definition of the candidate set C = C 1 ∪ C 2 and for an active resource r * , we have P 1 (t 1 |r * ) for t 1 ∈ C 1 and P 2 (t 2 |r * ) for t 2 ∈ C 2 . Then the set of recommended tags will be: T u, r ≔ argmax t∈C n (P 1 (t|r * )+ P 2 (t|r * )) where n is the number of recommended tags.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">ACT model</head><p>The model we use is called Author-Conference-Topic (ACT) model <ref type="bibr">[7]</ref>, which is an adaptation of topic model. The initial model was used to simultaneously modeling papers, authors, and publication venues within a unified probabilistic topic model. The model utilizes the topic distribution to represent the inter-dependencies among authors, papers and publication venues. In our task, for each 𝐰 d , 𝐭 d , u d ∈ D, we map the 𝐰 d to conference information, map 𝐭 d to author of the paper, map u d to the publication venue. Consequently, after modeling, we can get probabilistic relations among description, tags and user given a post. In details, we can get these interdependencies: P(z|t), P(w|z) and P(u|z), where z is a hidden topic layer in the topic model. From this model, we can utilize the hidden topic layer to learn some conceptual knowledge of the post. The number of topic z can be set manually. After obtaining the probability of P(z|t), P(w|z) and P(u|z) from the model, we can derive that for each word w ∈ 𝐰 d , each tag t ∈ 𝐭 d , we have:</p><p>To recommend tags for a given user u' and a given resource r', we will score all the tags t ∈ T using probabilistic methods as follows:</p><p>(3) There are two things that needed to be mentioned here. First, if the given user u ′ ∉ U, which means the given user is a new user. Then we cannot obtain P(u'|t) in the equation <ref type="bibr">(3)</ref>, in that case we will set the value to 1 and the equation (3) will become: </p><p>wr P t u r P t u r P t P u t P r t P t P u t P w t</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Results and analysis</head><p>As performance measures we use precision, recall and f-measure. For a given user u and a given resource r, the true tags are defined as TAG(u,r), then the precision, recall and f-measure of the recommended tags T (u, r) are defined as follows:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.1">Performance of Language model</head><p>In table <ref type="table">3</ref>, we show the performance of Language model on the test data provided by the organizers of ECML PKDD challenge 2009. We show the performance of bookmark, bibtex and the whole data respectively. From the table, we can find that the result of bookmark is better than that of bibtex and we can have a highest fmeasure of 13.949% when the number of recommended tags is  <ref type="table">3</ref>: performance of language model on the test data, the numbers are shown in the following format: recall/precision/f-measure</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.2">Performance of ACT model</head><p>In table <ref type="table">4</ref>, we show the performance of Language model on the test data provided by the organizers of ECML PKDD challenge 2009. We show the performance of bookmark, bibtex and the whole data respectively. From the table, we can see that the performance of ACT model is worse than the language model. We have the highest fmeasure of 3.077% when the number of recommended tags is set to five. Also, the ( , ) { | ( , , ')</p><p>The second step is to score all the tags in the candidate set and recommend the tags with the highest scores. In our proposed tag recommendation system, we score the tags in the candidate set using the following equation for all tags t∈ C(u, r):</p><p>(</p><p>where n(t,r) is the number of times that tag t has been posted to the resource r and n(t,R) is the number of times that tag t has been posted in the training data. T r is the number of tags posted to the resource r and λ is the Dirichlet smoothing factor and is commonly set according to the average document length, i.e. T /|R| In order to take the users' similarities into consideration, we change the equation (1) to the following equation: <ref type="bibr">(2)</ref> where U ′ = {u|(r, 𝐭, u) ∈ D} for a given resource r, m(t,u,r) is the number of times tag t has been posted to the resource r by the user u. The similarity of users sim(u,u') is define as follows,</p><p>For a given user u and a given resource r, the set of recommended tags will be:</p><p>T u, r : = arg t∈C u,r n P(t|r, u) where n is the number of recommended tags.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">FolkRank algorithm</head><p>FolkRank is a graph-based algorithm whose basic idea is to rank all the tags and pick out tags which are relatively important given a user u and a resource r. This algorithm is derived from the PageRank algorithm, which is used by the Google Internet search engine that assigns a numerical weighting to each element of a hyperlinked set of documents. The purpose of PageRank is to measure the hyperlink's relative importance within the set. However, due to the structural differences between hyperlinks and our tag recommendation system, we cannot apply the PageRank to our tag recommendation system and a new FolkRank algorithm was introduced in <ref type="bibr">[4,</ref><ref type="bibr">5]</ref>.</p><p>In order to apply a weight-spreading ranking scheme to recommend tags, we need to change the directed graph in PageRank to an undirected graph and change the corresponding ranking approach.</p><p>First, we convert the training dataset D into an undirected graph G = (V, E). V is the set of the nodes in the graph, which is composed of all the tags, resources and users in the training file, i.e. V = T ∪ R ∪ U. E is the set of the edges in the graph, which is defined as the co-occurrences of tags and users, users and resources, tags and</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Combination</head><p>We have proposed two different but similar methods for our tag recommendation system. Both are suitable to our case that the test data have already appeared in the training file, both make use of the similarity of users and resources, but the first method focuses more on the collaborative information while the second one focus more on the graph nodes and can spread the weight according to the co-occurrences. We hope to combine these two methods and get a better result.</p><p>We have tried some different approaches to combine these two methods. A simple method of combination is to multiply the scores of these two models and recommend tags with highest scores after combination. Details can be found in Algorithm 2.</p><p>Algorithm 2: the combination method used in our tag recommendation system 4 Experimental Results</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Dataset</head><p>We evaluate our experimental results using the evaluation methods provided by the organizers of ECML PKDD discovery challenge 2009.  <ref type="table">4</ref>: performance of two methods and combination on the test data, the numbers are shown in the following format: recall/precision/f-measure</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.3">Performance of combination</head><p>In table <ref type="table">4</ref>, we show the performance after the combination of the previous two methods. We are glad to see that the results after combination outperform these two methods. We have a 2% increase compared to the first method and a 4% increase compared to the second method. We have a highest f-measure of 32.622% when recommending 10 tags. The precision-recall plot in Fig. <ref type="figure">1</ref> reveals the quality of our recommendation system. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Conclusions</head><p>In this paper, we describe our tag recommendation system for the second task in the ECML PKDD Challenge 2009. We exploit two different methods to recommend tags when tags, resources, users in the test data are also in the training file. The experimental results show that the combination of these two methods will gain a better result. We need to further analyze the results to see which kind of information in the graph contributes more to the final ranking. Also, we can try to change the scoring scheme or expand the candidate set in our collaborative filtering method. Future work also includes some adaptations of PageRank for the tag recommendation system.</p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Enhanced hypertext categorization using hyperlinks</title>
		<author>
			<persName><forename type="first">Soumen</forename><surname>Chakrabarti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Byron</forename><forename type="middle">E</forename><surname>Dom</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Piotr</forename><surname>Indyk</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of SIGMOD-98</title>
				<editor>
			<persName><forename type="first">Laura</forename><forename type="middle">M</forename><surname>Haas</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Ashutosh</forename><surname>Tiwary</surname></persName>
		</editor>
		<meeting>SIGMOD-98<address><addrLine>New York, US</addrLine></address></meeting>
		<imprint>
			<publisher>ACM Press</publisher>
			<date type="published" when="1998">1998</date>
			<biblScope unit="page" from="307" to="318" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Machine-learning research: Four current directions</title>
		<author>
			<persName><forename type="first">G</forename><surname>Thomas</surname></persName>
		</author>
		<author>
			<persName><surname>Dietterich</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">The AI Magazine</title>
		<imprint>
			<biblScope unit="volume">18</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="97" to="136" />
			<date type="published" when="1998">1998</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Tag recommendations in social bookmarking systems</title>
		<author>
			<persName><forename type="first">Robert</forename><surname>Jaeschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Leandro</forename><surname>Marinho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Andreas</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lars</forename><surname>Schmidt-Thieme</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gerd</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">AI Communications</title>
		<imprint>
			<biblScope unit="page" from="231" to="247" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Link-based classification using labeled and unlabeled data. icml 2003 workshop on the continuum from labeled to unlabeled data</title>
		<author>
			<persName><forename type="first">Qing</forename><surname>Lu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lise</forename><surname>Getoor</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Machine Learning and Data Mining</title>
				<imprint>
			<date type="published" when="2003">2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Toward the ext Generation of Recommender. Systems: A Survey of the State-of-the-Art and Possible Extensions</title>
		<author>
			<persName><forename type="first">G</forename><surname>Adomavicius</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Tuzhilin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Knowledge and Data Engineering</title>
		<imprint>
			<biblScope unit="page" from="734" to="749" />
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">The Wraetlic LP Suite</title>
		<author>
			<persName><forename type="first">E</forename><surname>Alfonseca</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Moreno-Sandoval</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Guirao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ruiz-Casado</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 5th International Conference on Language Resources and Evaluation (LREC</title>
				<meeting>the 5th International Conference on Language Resources and Evaluation (LREC</meeting>
		<imprint>
			<date type="published" when="2006">2006. 2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Personalized Tag Recommendations via Tagging and Content-based Similarity Metrics</title>
		<author>
			<persName><forename type="first">A</forename><surname>Byde</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Cayzer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2007 International Conference on Weblogs and Social Media</title>
				<meeting>the 2007 International Conference on Weblogs and Social Media</meeting>
		<imprint>
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">P-TAG: Large Scale Automatic Generation of Personalized Annotation TAGs for the Web</title>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">A</forename><surname>Chirita</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Costache</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Handschuh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Nejdl</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 16th International Conference on World Wide Web (WWW</title>
				<meeting>the 16th International Conference on World Wide Web (WWW</meeting>
		<imprint>
			<date type="published" when="2007">2007. 2007</date>
			<biblScope unit="page" from="845" to="854" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Social Tag Prediction</title>
		<author>
			<persName><forename type="first">P</forename><surname>Heymann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Ramage</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Garcia-Molina</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 31st Annual International ACM Conference on Research and Development in Information Retrieval (SIGIR 2008)</title>
				<meeting>the 31st Annual International ACM Conference on Research and Development in Information Retrieval (SIGIR 2008)</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="531" to="538" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Information Retrieval in Folksonomies</title>
		<author>
			<persName><forename type="first">A</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Jäschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Schmitz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 3rd European Semantic Web Conference</title>
				<meeting>the 3rd European Semantic Web Conference<address><addrLine>ESWC</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2006">2006. 2006. 2006</date>
			<biblScope unit="page" from="411" to="426" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Tag Recommendations in Social Bookmarking Systems</title>
		<author>
			<persName><forename type="first">R</forename><surname>Jäschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Marinho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Schmidt-Thieme</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">AI Communications</title>
		<imprint>
			<biblScope unit="volume">21</biblScope>
			<biblScope unit="page" from="231" to="247" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Authoritative Sources in a Hyperlinked Environment</title>
		<author>
			<persName><forename type="first">J</forename><surname>Kleinberg</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of the ACM</title>
		<imprint>
			<biblScope unit="volume">46</biblScope>
			<biblScope unit="issue">5</biblScope>
			<biblScope unit="page" from="604" to="632" />
			<date type="published" when="1999">1999</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Two-way Poisson Mixture Models for Simultaneous Document Classification and Word Clustering</title>
		<author>
			<persName><forename type="first">J</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Zha</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computational Statistics &amp; Data Analysis</title>
		<imprint>
			<biblScope unit="volume">50</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="163" to="180" />
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">AutoTag: A Collaborative Approach to Automated Tag Assignment for Weblog Posts</title>
		<author>
			<persName><forename type="first">G</forename><surname>Mishne</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 15th International Conference on World Wide Web (WWW</title>
				<meeting>the 15th International Conference on World Wide Web (WWW</meeting>
		<imprint>
			<date type="published" when="2006">2006. 2006</date>
			<biblScope unit="page" from="953" to="954" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">A measure of Betweenness Centrality based on Random Walks</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">E J</forename><surname>Newman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Social Networks</title>
		<imprint>
			<biblScope unit="volume">27</biblScope>
			<biblScope unit="page" from="39" to="54" />
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<title level="m" type="main">The PageRank Citation Ranking: Bringing Order to the Web</title>
		<author>
			<persName><forename type="first">L</forename><surname>Page</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Brin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Motwani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Winograd</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1999">1999</date>
		</imprint>
		<respStmt>
			<orgName>Stanford InfoLab</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Technical Report</note>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">GroupLens: An Open Architecture for Collaborative Filtering of etnews</title>
		<author>
			<persName><forename type="first">P</forename><surname>Resnick</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Iacovou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Suchak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Bergstrom</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Riedl</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 1994 ACM conference on Computer Supported Cooperative Work (CSCW</title>
				<meeting>the 1994 ACM conference on Computer Supported Cooperative Work (CSCW</meeting>
		<imprint>
			<date type="published" when="1994">1994. 1994. 1994</date>
			<biblScope unit="page" from="175" to="186" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<title level="m" type="main">Introduction to Modern Information Retrieval</title>
		<author>
			<persName><forename type="first">G</forename><surname>Salton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">J</forename><surname>Mcgill</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1983">1983</date>
			<publisher>McGraw-Hill, Inc</publisher>
			<pubPlace>New York</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Real-time Automatic Tag Recommendation</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhuang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">C</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">L</forename><surname>Giles</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 31st Annual International ACM Conference on Research and Development in Information Retrieval (SIGIR 2008)</title>
				<meeting>the 31st Annual International ACM Conference on Research and Development in Information Retrieval (SIGIR 2008)</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="515" to="522" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">RSDC&apos;08: Tag Recommendations using Bookmark Content</title>
		<author>
			<persName><forename type="first">M</forename><surname>Tatu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Srikanth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>D'silva</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the ECML PKDD 2008 Discovery Challenge</title>
				<meeting>the ECML PKDD 2008 Discovery Challenge<address><addrLine>RSDC</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2008">2008. 2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Towards the Semantic Web: Collaborative Tag Suggestions</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Fu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Mao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Su</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the WWWW 2006 Workshop on Collaborative Web Tagging</title>
				<meeting>of the WWWW 2006 Workshop on Collaborative Web Tagging</meeting>
		<imprint>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Bipartite Graph Partitioning and Data Clustering</title>
		<author>
			<persName><forename type="first">H</forename><surname>Zha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Ding</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Simon</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 10th ACM International Conference on Information and Knowledge (CIKM 2001)</title>
				<meeting>the 10th ACM International Conference on Information and Knowledge (CIKM 2001)</meeting>
		<imprint>
			<date type="published" when="2001">2001</date>
			<biblScope unit="page" from="25" to="32" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Tag Recommendation for Folksonomies Oriented towards Individual Users</title>
		<author>
			<persName><forename type="first">Marek</forename><surname>Lipczak</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ECML</title>
		<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">RSDC&apos;08: Tag Recommendations using Bookmark Content</title>
		<author>
			<persName><forename type="first">Marta</forename><surname>Tatu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Munirathnam</forename><surname>Srikanth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Thomas D'</forename><surname>Silva</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of ECML PKDD Discovery Challenge 2008</title>
				<meeting>ECML PKDD Discovery Challenge 2008<address><addrLine>RSDC</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Multilabel Text Classifcation for Automated Tag Suggestion</title>
		<author>
			<persName><forename type="first">Ioannis</forename><surname>Katakis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Grigorios</forename><surname>Tsoumakas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ioannis</forename><surname>Vlahavas</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of ECML PKDD Discovery Challenge</title>
				<meeting>ECML PKDD Discovery Challenge<address><addrLine>RSDC</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2008">2008. 2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<monogr>
		<author>
			<persName><forename type="first">Paul</forename><surname>Heymann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Daniel</forename><surname>Ramage</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hector</forename><surname>Garcia-Molina</surname></persName>
		</author>
		<title level="m">Social Tag Prediction, SIGIR&apos;08</title>
				<meeting><address><addrLine>Singapore</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2008">July 20-24, 2008</date>
			<biblScope unit="page" from="531" to="538" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">AutoTag: A Collaborative Approach to Automated Tag Assignment for Weblog Posts</title>
		<author>
			<persName><forename type="first">Gilad</forename><surname>Mishne</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">WWW</title>
		<imprint>
			<biblScope unit="page" from="953" to="954" />
			<date type="published" when="2006-05-22">2006. May 22-26, 2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">Tag Recommendations in Folksonomies</title>
		<author>
			<persName><forename type="first">Robert</forename><surname>J¨aschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Andreas</forename><surname>Leandromarinho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lars</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gerd</forename><surname>Schmidt-Thieme</surname></persName>
		</author>
		<author>
			<persName><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">PKDD 2007</title>
				<editor>
			<persName><forename type="first">J</forename><forename type="middle">N</forename><surname>Kok</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="2007">2007</date>
			<biblScope unit="volume">4702</biblScope>
			<biblScope unit="page" from="506" to="514" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<monogr>
		<author>
			<persName><forename type="first">Adriana</forename><surname>Budura</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Sebastian</forename><surname>Michel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Philippe</forename><surname>Cudre-Mauroux</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Karl</forename><surname>Aberer</surname></persName>
		</author>
		<title level="m">Neighborhood-based Tag Prediction, 6th Annual European Semantic Web Conference (ESWC2009)</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<analytic>
		<title level="a" type="main">Large margin rank boundaries for ordinal regression</title>
		<author>
			<persName><surname>Herbrich</surname></persName>
		</author>
		<author>
			<persName><surname>Graepel</surname></persName>
		</author>
		<author>
			<persName><surname>Obermayer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the eighth ACM SIGKDD international</title>
				<meeting>the eighth ACM SIGKDD international</meeting>
		<imprint>
			<biblScope unit="volume">02</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<analytic>
		<title level="a" type="main">Adapting Ranking SVM to Document Retrieval</title>
		<author>
			<persName><forename type="first">Cao</forename><surname>Yunbo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><forename type="middle">U</forename><surname>Jun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Liu</forename><surname>Tie-Yan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">I</forename><surname>Hang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Huang</forename><surname>Yalou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hon</forename><surname>Hsiao-Wuen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">SIGIR&apos;06</title>
				<meeting><address><addrLine>Seattle; Washington, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2006">August 6-11,2006</date>
			<biblScope unit="page" from="186" to="193" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b31">
	<analytic>
		<title level="a" type="main">On some clustering algorithms for document maps creation</title>
		<author>
			<persName><forename type="first">Draminski</forename><surname>Ciesielski</surname></persName>
		</author>
		<author>
			<persName><surname>Klopotek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Wierzchon</forename><surname>Kujawiak</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Intelligent Information Processing and Web Mining Conference (IIS:IIPWM-2005)</title>
				<meeting>the Intelligent Information Processing and Web Mining Conference (IIS:IIPWM-2005)<address><addrLine>Gdansk,</addrLine></address></meeting>
		<imprint>
			<publisher>Springer-Verlag</publisher>
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
	<note>Advances in Soft Computing</note>
</biblStruct>

<biblStruct xml:id="b32">
	<analytic>
		<title level="a" type="main">Tag Recommendations in Social Bookmarking Systems</title>
		<author>
			<persName><forename type="first">Marinho</forename><surname>Jäschke</surname></persName>
		</author>
		<author>
			<persName><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Stumme</forename><surname>Schmidt-Thieme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">AI Communications</title>
				<meeting><address><addrLine>Amsterdam</addrLine></address></meeting>
		<imprint>
			<publisher>IOS Press</publisher>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b33">
	<analytic>
		<title level="a" type="main">Multilabel Text Classification for Automated Tag Suggestion</title>
		<author>
			<persName><forename type="first">Tsoumakas</forename><surname>Katakis</surname></persName>
		</author>
		<author>
			<persName><surname>Vlahavas</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of ECML PKDD Discovery Challenge (RSDC08)</title>
				<meeting>ECML PKDD Discovery Challenge (RSDC08)</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b34">
	<analytic>
		<title level="a" type="main">Coexistence of Fuzzy and Crisp Concepts in Document Maps</title>
		<author>
			<persName><forename type="first">Wierzchon</forename><surname>Klopotek</surname></persName>
		</author>
		<author>
			<persName><surname>Ciesielski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kujawiak</forename><surname>Draminski</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the International Conference on Artificial Neural Networks (ICANN 2005)</title>
		<title level="s">Lecture Notes in Artificial Intelligence</title>
		<meeting>the International Conference on Artificial Neural Networks (ICANN 2005)</meeting>
		<imprint>
			<publisher>Springer-Verlag</publisher>
			<date type="published" when="2005">2005</date>
			<biblScope unit="volume">3697</biblScope>
		</imprint>
	</monogr>
	<note>LNAI</note>
</biblStruct>

<biblStruct xml:id="b35">
	<analytic>
		<title level="a" type="main">Tag Recommendation for Folksonomies Oriented towards Individual Users</title>
		<author>
			<persName><surname>Lipczak</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of ECML PKDD Discovery Challenge (RSDC08)</title>
				<meeting>ECML PKDD Discovery Challenge (RSDC08)</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b36">
	<monogr>
		<title level="m" type="main">Wikinomics</title>
		<author>
			<persName><forename type="first">Williams</forename><surname>Tapscott</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2008">2008</date>
			<publisher>Atlantic Books</publisher>
			<pubPlace>London</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b37">
	<analytic>
		<title level="a" type="main">Tag Recommendations using Bookmark Content</title>
		<author>
			<persName><forename type="first">Srikanth</forename><surname>Tatu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D'</forename><surname>Silva</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of ECML PKDD Discovery Challenge (RSDC08)</title>
				<meeting>ECML PKDD Discovery Challenge (RSDC08)</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
	<note>RSDC&apos;08</note>
</biblStruct>

<biblStruct xml:id="b38">
	<analytic>
		<title level="a" type="main">Contag: A semantic tag recommendation system</title>
		<author>
			<persName><forename type="first">B</forename><surname>Adrian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Sauermann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Roth-Berghofer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of I-Semantics&apos; 07</title>
				<editor>
			<persName><forename type="first">T</forename><surname>Pellegrini</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Schaffert</surname></persName>
		</editor>
		<meeting>I-Semantics&apos; 07</meeting>
		<imprint>
			<publisher>JUCS</publisher>
			<date type="published" when="2007">2007</date>
			<biblScope unit="page" from="297" to="304" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b39">
	<analytic>
		<title level="a" type="main">Recommending smart tags in a social bookmarking system</title>
		<author>
			<persName><forename type="first">P</forename><surname>Basile</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Gendarmi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Lanubile</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Semeraro</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Bridging the Gep between Semantic Web and Web 2</title>
				<meeting><address><addrLine>SemNet</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2007">2007. 2007</date>
			<biblScope unit="page" from="22" to="29" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b40">
	<monogr>
		<title level="m" type="main">Generalized cores</title>
		<author>
			<persName><forename type="first">V</forename><surname>Batagelj</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Zaveršnik</surname></persName>
		</author>
		<idno>preprint cs/0202039</idno>
		<imprint>
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
	<note type="report_type">Arxiv</note>
</biblStruct>

<biblStruct xml:id="b41">
	<analytic>
		<title level="a" type="main">Usage patterns of collaborative tagging systems</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">A</forename><surname>Golder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">A</forename><surname>Huberman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Information Science</title>
		<imprint>
			<biblScope unit="volume">32</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page">198</biblScope>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b42">
	<analytic>
		<title level="a" type="main">Social tag prediction</title>
		<author>
			<persName><forename type="first">P</forename><surname>Heymann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Ramage</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Garcia-Molina</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">SIGIR &apos;08: Proceedings of the 31st annual international ACM SIGIR conference on Research and development in information retrieval</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="531" to="538" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b43">
	<analytic>
		<title level="a" type="main">Information retrieval in folksonomies: Search and ranking</title>
		<author>
			<persName><forename type="first">A</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Jaschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Schmitz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Lecture Notes in Computer Science</title>
		<imprint>
			<biblScope unit="volume">4011</biblScope>
			<biblScope unit="page">411</biblScope>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b44">
	<analytic>
		<title level="a" type="main">Tag Recommendations in Folksonomies</title>
		<author>
			<persName><forename type="first">R</forename><surname>Jaschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Marinho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Schmidt-Thieme</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">LECTURE NOTES IN COMPUTER SCIENCE</title>
		<imprint>
			<biblScope unit="volume">4702</biblScope>
			<biblScope unit="page">506</biblScope>
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b45">
	<analytic>
		<title level="a" type="main">Tag recommendation for folksonomies oriented towards individual users</title>
		<author>
			<persName><forename type="first">M</forename><surname>Lipczak</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the ECML/PKDD 2008 Discovery Challenge Workshop, part of the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases</title>
				<meeting>the ECML/PKDD 2008 Discovery Challenge Workshop, part of the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b46">
	<analytic>
		<title level="a" type="main">Collaborative tagging as a knowledge organisation and resource discovery tool</title>
		<author>
			<persName><forename type="first">G</forename><surname>Macgregor</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Mcculloch</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Library Review</title>
		<imprint>
			<biblScope unit="volume">55</biblScope>
			<biblScope unit="issue">5</biblScope>
			<biblScope unit="page" from="291" to="300" />
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b47">
	<analytic>
		<title level="a" type="main">Folksonomies-Cooperative Classification and Communication Through Shared Metadata</title>
		<author>
			<persName><forename type="first">A</forename><surname>Mathes</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Computer Mediated Communication</title>
				<imprint>
			<date type="published" when="2004-12">December, 2004</date>
		</imprint>
		<respStmt>
			<orgName>Graduate School of Library and Information Science, University of Illinois Urbana-Champaign</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Doctoral Seminar</note>
</biblStruct>

<biblStruct xml:id="b48">
	<analytic>
		<title level="a" type="main">Ontologies are us: A unified model of social networks and semantics</title>
		<author>
			<persName><forename type="first">P</forename><surname>Mika</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Web Semantics: Science, Services and Agents on the World Wide Web</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="5" to="15" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b49">
	<analytic>
		<title level="a" type="main">Investigation of the effectiveness of tag-based contextual collaborative filtering in website recommendation</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">Y</forename><surname>Nakamoto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Nakajima</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Miyazaki</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Uemura</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Kato</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Communication Systems and Electrical Engineering</title>
				<imprint>
			<publisher>Springerlink</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="309" to="318" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b50">
	<analytic>
		<title level="a" type="main">Reasonable tag-based collaborative filtering for social tagging systems</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">Y</forename><surname>Nakamoto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Nakajima</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Miyazaki</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Uemura</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Kato</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Inagaki</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">WICOW &apos;08: Proceeding of the 2nd ACM workshop on Information credibility on the web</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="11" to="18" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b51">
	<analytic>
		<title level="a" type="main">Mining association rules in folksonomies</title>
		<author>
			<persName><forename type="first">C</forename><surname>Schmitz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Jaschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. IFCS 2006 Conference</title>
				<meeting>IFCS 2006 Conference</meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="261" to="270" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b52">
	<analytic>
		<title level="a" type="main">Flickr tag recommendation based on collective knowledge</title>
		<author>
			<persName><forename type="first">B</forename><surname>Sigurbjörnsson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Van Zwol</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">WWW &apos;08: Proceeding of the 17th international conference on World Wide Web</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="327" to="336" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b53">
	<analytic>
		<title level="a" type="main">Realtime automatic tag recommendation</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhuang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W.-C</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">L</forename><surname>Giles</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">SIGIR &apos;08: Proceedings of the 31st annual international ACM SIGIR conference on Research and development in information retrieval</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="515" to="522" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b54">
	<analytic>
		<title level="a" type="main">Tag-aware recommender systems by fusion of collaborative filtering algorithms</title>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">H L</forename><surname>Tso-Sutter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">B</forename><surname>Marinho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Schmidt-Thieme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">SAC &apos;08: Proceedings of the 2008 ACM symposium on Applied computing</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="1995" to="1999" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b55">
	<analytic>
		<title level="a" type="main">Towards the semantic web: Collaborative tag suggestions</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Fu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Mao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Su</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Collaborative Web Tagging Workshop at WWW2006</title>
				<meeting><address><addrLine>Edinburgh, Scotland</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2006-05">May, 2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b56">
	<analytic>
		<title level="a" type="main">An algorithm for text categorization</title>
		<author>
			<persName><forename type="first">A</forename><surname>Gkanogiannis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Kalampoukis</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ACM International Conference on Research and Development in Information Retrieval SIGIR-2008</title>
				<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="869" to="870" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b57">
	<analytic>
		<title level="a" type="main">A novel supervised learning algorithm and its use for spam detection in social bookmarking systems</title>
		<author>
			<persName><forename type="first">A</forename><surname>Gkanogiannis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Kalamboukis</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ECML PKDD Discovery Challenge &apos;08</title>
				<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b58">
	<analytic>
		<title level="a" type="main">The perceptron: a probabilistic model for information storage and organization in the brain</title>
		<author>
			<persName><forename type="first">F</forename><surname>Rosenblatt</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Psychological Review</title>
		<imprint>
			<biblScope unit="volume">65</biblScope>
			<biblScope unit="issue">6</biblScope>
			<biblScope unit="page" from="386" to="408" />
			<date type="published" when="1958-11">November 1958</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b59">
	<analytic>
		<title level="a" type="main">On convergence proofs for perceptrons</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">B</forename><surname>Novikoff</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Symposium on the Mathematical Theory of Automata</title>
				<meeting>the Symposium on the Mathematical Theory of Automata</meeting>
		<imprint>
			<date type="published" when="1963">1963</date>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="page" from="615" to="622" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b60">
	<monogr>
		<title level="m" type="main">A study on thresholding strategies for text categorization</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Yang</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2001">2001</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b61">
	<monogr>
		<title level="m" type="main">Weight adjustment schemes for a centroid based classifier</title>
		<author>
			<persName><forename type="first">G</forename><surname>Karypis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Shankar</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2000">2000</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b62">
	<monogr>
		<title level="m" type="main">Relevance feedback and other query modification techniques</title>
		<author>
			<persName><forename type="first">D</forename><surname>Harman</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1992">1992</date>
			<biblScope unit="page" from="241" to="263" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b63">
	<analytic>
		<title level="a" type="main">Optimization of relevance feedback weights</title>
		<author>
			<persName><forename type="first">C</forename><surname>Buckley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Salton</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">SIGIR &apos;95: Proceedings of the 18th annual international ACM SIGIR conference on Research and development in information retrieval</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="1995">1995</date>
			<biblScope unit="page" from="351" to="357" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b64">
	<analytic>
		<title level="a" type="main">An algorithm for suffix stripping</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">F</forename><surname>Porter</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Program</title>
		<imprint>
			<biblScope unit="volume">14</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="130" to="137" />
			<date type="published" when="1980">1980</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b65">
	<monogr>
		<title level="m" type="main">TreeTagger is run on the vocabulary file once for each language: English, French, German, Italian, Spanish and Dutch</title>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b66">
	<monogr>
		<ptr target="http://arxiv.org/abs/cs.DL/0508082" />
		<title level="m">The structure of collaborative tagging systems</title>
				<editor>
			<persName><forename type="first">S</forename><surname>Golder</surname></persName>
		</editor>
		<editor>
			<persName><surname>Huberman</surname></persName>
		</editor>
		<meeting><address><addrLine>, B.A.</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2005-08">Aug 2005</date>
		</imprint>
	</monogr>
	<note>Using this &quot;&lt;unknown&gt;&quot; word, we combine the output of all six lemmatized files. If a term is not recognized by any language, the term itself is used as lemma</note>
</biblStruct>

<biblStruct xml:id="b67">
	<monogr>
		<title level="m" type="main">Ecml pkdd discovery challenge</title>
		<author>
			<persName><forename type="first">F</forename><surname>Eisterlehner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Jäschke</surname></persName>
		</author>
		<ptr target="http://www.kde.cs.uni-kassel.de/ws/dc09" />
		<imprint>
			<date type="published" when="2009">2009. 2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b68">
	<analytic>
		<title level="a" type="main">Flickr tag recommendation based on collective knowledge</title>
		<author>
			<persName><forename type="first">B</forename><surname>Sigurbjörnsson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Van Zwol</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">WWW &apos;08: Proceeding of the 17th international conference on World Wide Web</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="327" to="336" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b69">
	<monogr>
		<title level="m" type="main">Tag recommendations in folksonomies</title>
		<author>
			<persName><forename type="first">R</forename><surname>Jäschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Marinho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Schmidt-Thieme</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Stumme</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2007">2007</date>
			<biblScope unit="page" from="506" to="514" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b70">
	<analytic>
		<title level="a" type="main">Tag recommendations in social bookmarking systems</title>
		<author>
			<persName><forename type="first">R</forename><surname>Jschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Marinho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Schmidt-Thieme</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">AI Communications</title>
		<imprint>
			<biblScope unit="volume">21</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="231" to="247" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b71">
	<analytic>
		<title level="a" type="main">Tag recommendations based on tensor dimensionality reduction</title>
		<author>
			<persName><forename type="first">P</forename><surname>Symeonidis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Nanopoulos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Manolopoulos</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">RecSys &apos;08: Proceedings of the 2008 ACM conference on Recommender systems</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="43" to="50" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b72">
	<analytic>
		<title level="a" type="main">Tag recommendation for folksonomies oriented towards individual users</title>
		<author>
			<persName><forename type="first">M</forename><surname>Lipczak</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ECML PKDD Discovery Challenge</title>
				<imprint>
			<date type="published" when="2008">2008. 2008</date>
			<biblScope unit="page" from="84" to="95" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b73">
	<analytic>
		<title level="a" type="main">Rsdc 08: Tag recommendation using bookmark content</title>
		<author>
			<persName><forename type="first">M</forename><surname>Tatu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Srikanth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>D'silva</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ECML PKDD Discovery Challenge</title>
				<imprint>
			<date type="published" when="2008">2008. 2008</date>
			<biblScope unit="page" from="96" to="107" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b74">
	<monogr>
		<title level="m" type="main">Recent developments in document clustering</title>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">O</forename><surname>Andrews</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">A</forename><surname>Fox</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2007">2007</date>
		</imprint>
		<respStmt>
			<orgName>Computer Science, Virginia Tech</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Technical report</note>
</biblStruct>

<biblStruct xml:id="b75">
	<analytic>
		<title level="a" type="main">A Survey of Clustering Data Mining Techniques</title>
		<author>
			<persName><forename type="first">J</forename><surname>Kogan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Nicholas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Teboulle</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Grouping Multidimensional Data</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="25" to="71" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b76">
	<analytic>
		<title level="a" type="main">Automated tag clustering: Improving search and exploration in the tag space</title>
		<author>
			<persName><forename type="first">G</forename><surname>Begelman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the Collaborative Web Tagging Workshop at WWW06</title>
				<meeting>of the Collaborative Web Tagging Workshop at WWW06</meeting>
		<imprint>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b77">
	<analytic>
		<title level="a" type="main">Personalized recommendation in social tagging systems using hierarchical clustering</title>
		<author>
			<persName><forename type="first">A</forename><surname>Shepitsen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gemmell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Mobasher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Burke</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">RecSys &apos;08: Proceedings of the 2008 ACM conference on Recommender systems</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="259" to="266" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b78">
	<monogr>
		<ptr target="http://www.bibsonomy.org/" />
		<title level="m">Bibsonomy: A blue social bookmark and publication sharing system</title>
				<imprint>
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b79">
	<analytic>
		<title level="a" type="main">A robust discriminative term weighting based linear discriminant method for text classification</title>
		<author>
			<persName><forename type="first">K</forename><surname>Junejo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Karim</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ICDM &apos;08. Eighth IEEE International Conference on</title>
				<imprint>
			<date type="published" when="2008-12">2008. Dec. 2008</date>
			<biblScope unit="page" from="323" to="332" />
		</imprint>
	</monogr>
	<note>Data Mining</note>
</biblStruct>

<biblStruct xml:id="b80">
	<monogr>
		<ptr target="http://www.citeulike.org/" />
		<title level="m">Citeulike website</title>
				<imprint>
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
	<note>CiteULike</note>
</biblStruct>

<biblStruct xml:id="b81">
	<monogr>
		<ptr target="http://www.ims.uni-stuttgart.de/projekte/corplex/TreeTagger/" />
		<title level="m">TreeTagger: Treetagger -a language independent part-of-speech tagger</title>
				<imprint>
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b82">
	<analytic>
		<title level="a" type="main">L3s at inex 2008: Retrieving entities using structured information</title>
		<author>
			<persName><forename type="first">Nick</forename><surname>Craswell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gianluca</forename><surname>Demartini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Julien</forename><surname>Gaugaz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Tereza</forename><surname>Iofciu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">INEX</title>
				<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b83">
	<analytic>
		<title level="a" type="main">Overview of the inex 2008 entity ranking track</title>
		<author>
			<persName><forename type="first">Gianluca</forename><surname>Demartini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Arjen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Tereza</forename><surname>De Vries</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jianhan</forename><surname>Iofciu</surname></persName>
		</author>
		<author>
			<persName><surname>Zhu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">INEX</title>
				<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b84">
	<analytic>
		<title level="a" type="main">Tag recommendations in social bookmarking systems</title>
		<author>
			<persName><forename type="first">Robert</forename><surname>Jäschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Leandro</forename><forename type="middle">Balby</forename><surname>Marinho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Andreas</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lars</forename><surname>Schmidt-Thieme</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gerd</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">AI Commun</title>
		<imprint>
			<biblScope unit="volume">21</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="231" to="247" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b85">
	<analytic>
		<title level="a" type="main">Tag Recommendation for Folksonomies Oriented towards Individual Users</title>
		<author>
			<persName><forename type="first">Marek</forename><surname>Lipczak</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the ECML/PKDD 2008 Discovery Challenge Workshop, part of the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases</title>
				<meeting>the ECML/PKDD 2008 Discovery Challenge Workshop, part of the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b86">
	<analytic>
		<title level="a" type="main">RSDC&apos;08: Tag Recommendations using Bookmark Content</title>
		<author>
			<persName><forename type="first">Marta</forename><surname>Tatu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Munirathnam</forename><surname>Srikanth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Thomas D'</forename><surname>Silva</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the ECML/PKDD 2008 Discovery Challenge Workshop, part of the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases</title>
				<meeting>the ECML/PKDD 2008 Discovery Challenge Workshop, part of the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b87">
	<analytic>
		<title level="a" type="main">What is web 2.0: Design patterns and business models for the next generation of software</title>
		<author>
			<persName><forename type="first">T</forename><surname>Oreilly</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Social Science Research Network Working Paper Series</title>
		<imprint>
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b88">
	<analytic>
		<title level="a" type="main">Rsdc&apos;08: Tag recommendations using bookmark content</title>
		<author>
			<persName><forename type="first">M</forename><surname>Tatu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Srikanth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>D'silva</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ECML PKDD Discovery Challenge</title>
				<imprint>
			<date type="published" when="2008">2008. 2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b89">
	<monogr>
		<title level="m" type="main">WordNet-An Electronic Lexical Database</title>
		<author>
			<persName><forename type="first">C</forename><surname>Fellbaum</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1998">1998</date>
			<publisher>MIT Press</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b90">
	<analytic>
		<title level="a" type="main">Multilabel text classification for automated tag suggestion</title>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">K G T I</forename><surname>Vlahavas</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ECML PKDD Discovery Challenge</title>
				<imprint>
			<date type="published" when="2008">2008. 2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b91">
	<analytic>
		<title level="a" type="main">Autotag: A collaborative approach to automated tag assignment to weblog posts</title>
		<author>
			<persName><forename type="first">G</forename><surname>Mishne</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of WWW&apos;06</title>
				<meeting>WWW&apos;06<address><addrLine>Edinburgh, Scotland</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b92">
	<analytic>
		<title level="a" type="main">New methods in automatic extracting</title>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">P</forename><surname>Edmundson</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of the ACM</title>
		<imprint>
			<biblScope unit="volume">16</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="264" to="285" />
			<date type="published" when="1969">1969</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b93">
	<analytic>
		<title level="a" type="main">A trainable document summarizer</title>
		<author>
			<persName><forename type="first">J</forename><surname>Kupiec</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Pedersen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Chen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">SIGIR &apos;95: Proceedings of the 18th annual international ACM SIGIR conference on Research and development in information retrieval</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="1995">1995</date>
			<biblScope unit="page" from="68" to="73" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b94">
	<monogr>
		<title level="m" type="main">CRFTagger: CRF English POS Tagger</title>
		<author>
			<persName><forename type="first">X</forename><surname>Phan</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b95">
	<monogr>
		<title level="m" type="main">Advances in Automatic Text Summarization</title>
		<editor>Mani, I., Maybury, M.T.</editor>
		<imprint>
			<date type="published" when="1998">1998</date>
			<publisher>The MIT Press</publisher>
			<pubPlace>Cambridge, MA</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b96">
	<analytic>
		<title level="a" type="main">Ht06, tagging paper, taxonomy, flickr, academic article</title>
		<author>
			<persName><forename type="first">C</forename><surname>Marlow</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Naaman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Boyd</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Davis</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 17th ACM Conference on Hypertext and Hypermedia</title>
				<editor>
			<persName><forename type="first">U</forename><forename type="middle">K</forename><surname>Wiil</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><forename type="middle">J</forename><surname>Nürnberg</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Rubart</surname></persName>
		</editor>
		<meeting>the 17th ACM Conference on Hypertext and Hypermedia<address><addrLine>Odense, Denmark, New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2006">August 22-25, 2006. 2006</date>
			<biblScope unit="page" from="31" to="40" />
		</imprint>
	</monogr>
	<note>HYPERTEXT 2006</note>
</biblStruct>

<biblStruct xml:id="b97">
	<analytic>
		<title level="a" type="main">Usage patterns of collaborative tagging systems</title>
		<author>
			<persName><forename type="first">S</forename><surname>Golder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">A</forename><surname>Huberman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Information Science</title>
		<imprint>
			<biblScope unit="volume">32</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="198" to="208" />
			<date type="published" when="2006-04">April 2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b98">
	<analytic>
		<title level="a" type="main">Social tag prediction</title>
		<author>
			<persName><forename type="first">P</forename><surname>Heymann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Ramage</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Garcia-Molina</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">SIGIR &apos;08: Proceedings of the 31st annual international ACM SIGIR conference on Research and development in information retrieval</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="531" to="538" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b99">
	<analytic>
		<title level="a" type="main">Latent dirichlet allocation</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">M</forename><surname>Blei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">Y</forename><surname>Ng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">I</forename><surname>Jordan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Machine Learning Research</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="page" from="993" to="1022" />
			<date type="published" when="2003-01">January 2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b100">
	<analytic>
		<title level="a" type="main">Finding scientific topics</title>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">L</forename><surname>Griffiths</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Steyvers</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Proc Natl Acad Sci U S A</title>
		<imprint>
			<biblScope unit="volume">101</biblScope>
			<biblScope unit="page" from="5228" to="5235" />
			<date type="published" when="2004-04">April 2004</date>
		</imprint>
	</monogr>
	<note>Suppl</note>
</biblStruct>

<biblStruct xml:id="b101">
	<analytic>
		<title level="a" type="main">A latent dirichlet model for unsupervised entity resolution</title>
		<author>
			<persName><forename type="first">I</forename><surname>Bhattacharya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Getoor</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">SIAM Conference on Data Mining (SDM)</title>
				<imprint>
			<date type="published" when="2006-04">April 2006</date>
			<biblScope unit="page" from="47" to="58" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b102">
	<analytic>
		<title level="a" type="main">Linked latent dirichlet allocation in web spam filtering</title>
		<author>
			<persName><forename type="first">I</forename><surname>Bíró</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Siklósi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Szabó</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">A</forename><surname>Benczúr</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">AIRWeb &apos;09: Proceedings of the 5th International Workshop on Adversarial Information Retrieval on the Web</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2009">2009</date>
			<biblScope unit="page" from="37" to="40" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b103">
	<monogr>
		<title level="m" type="main">Mallet: A machine learning for language toolkit</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">K</forename><surname>Mccallum</surname></persName>
		</author>
		<ptr target="http://mallet.cs.umass.edu" />
		<imprint>
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b104">
	<analytic>
		<title level="a" type="main">Efficient methods for topic model inference on streaming document collections</title>
		<author>
			<persName><forename type="first">L</forename><surname>Yao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Mimno</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mccallum</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">KDD &apos;09: Proceedings of the 15th ACM SIGKDD conference on Knowledge Discovery and Data Mining</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b105">
	<analytic>
		<title level="a" type="main">Autotag: a collaborative approach to automated tag assignment for weblog posts</title>
		<author>
			<persName><forename type="first">G</forename><surname>Mishne</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">WWW &apos;06: Proceedings of the 15th international conference on World Wide Web</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="953" to="954" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b106">
	<analytic>
		<title level="a" type="main">P-tag: large scale automatic generation of personalized annotation tags for the web</title>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">A</forename><surname>Chirita</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Costache</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Nejdl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Handschuh</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">WWW &apos;07: Proceedings of the 16th international conference on World Wide Web</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2007">2007</date>
			<biblScope unit="page" from="845" to="854" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b107">
	<analytic>
		<title level="a" type="main">Towards the semantic web: Collaborative tag suggestions</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Fu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Mao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Su</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of Collaborative Web Tagging Workshop at 15th International World Wide Web Conference</title>
				<meeting>Collaborative Web Tagging Workshop at 15th International World Wide Web Conference</meeting>
		<imprint>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b108">
	<analytic>
		<title level="a" type="main">Flickr tag recommendation based on collective knowledge</title>
		<author>
			<persName><forename type="first">B</forename><surname>Sigurbjörnsson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Van Zwol</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">WWW &apos;08: Proceeding of the 17th international conference on World Wide Web</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="327" to="336" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b109">
	<analytic>
		<title level="a" type="main">Tag recommendations in folksonomies</title>
		<author>
			<persName><forename type="first">R</forename><surname>Jäschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">B</forename><surname>Marinho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Schmidt-Thieme</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Knowledge Discovery in Databases: PKDD 2007, 11th European Conference on Principles and Practice of Knowledge Discovery in Databases</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">J</forename><forename type="middle">N</forename><surname>Kok</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Koronacki</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><forename type="middle">L</forename><surname>De Mántaras</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Matwin</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Mladenic</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Skowron</surname></persName>
		</editor>
		<meeting><address><addrLine>Warsaw, Poland; Heidelberg, Germany</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2007">September 17-21, 2007. 2007</date>
			<biblScope unit="volume">4702</biblScope>
			<biblScope unit="page" from="506" to="514" />
		</imprint>
	</monogr>
	<note>Proceedings.</note>
</biblStruct>

<biblStruct xml:id="b110">
	<analytic>
		<title level="a" type="main">Tag recommendations based on tensor dimensionality reduction</title>
		<author>
			<persName><forename type="first">P</forename><surname>Symeonidis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Nanopoulos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Manolopoulos</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">RecSys &apos;08: Proceedings of the 2008 ACM conference on Recommender systems</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="43" to="50" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b111">
	<analytic>
		<title level="a" type="main">Personalized, interactive tag recommendation for flickr</title>
		<author>
			<persName><forename type="first">N</forename><surname>Garg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Weber</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">RecSys &apos;08: Proceedings of the 2008 ACM conference on Recommender systems</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="67" to="74" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b112">
	<analytic>
		<title level="a" type="main">Personalized recommendation in social tagging systems using hierarchical clustering</title>
		<author>
			<persName><forename type="first">A</forename><surname>Shepitsen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gemmell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Mobasher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">D</forename><surname>Burke</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2008 ACM Conference on Recommender Systems, RecSys 2008</title>
				<editor>
			<persName><forename type="first">P</forename><surname>Pu</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><forename type="middle">G</forename><surname>Bridge</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Mobasher</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Ricci</surname></persName>
		</editor>
		<meeting>the 2008 ACM Conference on Recommender Systems, RecSys 2008<address><addrLine>Lausanne, Switzerland; New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2008">October 23-25, 2008. 2008</date>
			<biblScope unit="page" from="259" to="266" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b113">
	<analytic>
		<title level="a" type="main">Latent dirichlet allocation for tag recommendation</title>
		<author>
			<persName><forename type="first">R</forename><surname>Krestel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Fankhauser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Nejdl</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">3rd ACM Conference on Recommender Systems</title>
				<meeting><address><addrLine>New York City, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2009">October 22-25, 2009. 2009</date>
		</imprint>
	</monogr>
	<note>to appear</note>
</biblStruct>

<biblStruct xml:id="b114">
	<analytic>
		<title level="a" type="main">Real-time automatic tag recommendation</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhuang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">C</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">L</forename><surname>Giles</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">SIGIR &apos;08: Proceedings of the 31st annual international ACM SIGIR conference on Research and development in information retrieval</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="515" to="522" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b115">
	<analytic>
		<title level="a" type="main">Social Tag Prediction</title>
		<author>
			<persName><forename type="first">P</forename><surname>Heymann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Ramage</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Garcia-Molina</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 31st annual international ACM SIGIR conference on Research and development in information retrieval (SIGIR 2008)</title>
				<meeting>the 31st annual international ACM SIGIR conference on Research and development in information retrieval (SIGIR 2008)</meeting>
		<imprint>
			<biblScope unit="page" from="531" to="538" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b116">
	<analytic>
		<title level="a" type="main">RSDC&apos;08: Tag Recommendations using Bookmark Content</title>
		<author>
			<persName><forename type="first">M</forename><surname>Tatu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Srikanth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Silva</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of ECML PKDD Discovery Challenge 2008</title>
				<meeting>ECML PKDD Discovery Challenge 2008<address><addrLine>RSDC</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="96" to="107" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b117">
	<analytic>
		<title level="a" type="main">Tag Recommendation for Folksonomies Oriented towards Individual Users</title>
		<author>
			<persName><forename type="first">M</forename><surname>Lipczak</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of ECML PKDD Discovery Challenge (RSDC 2008)</title>
				<meeting>ECML PKDD Discovery Challenge (RSDC 2008)</meeting>
		<imprint>
			<biblScope unit="page" from="84" to="95" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b118">
	<analytic>
		<title level="a" type="main">A Language Modeling Approach to Information Retrieval</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Ponte</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W.-B</forename><surname>Croft</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 21st annual international ACM SIGIR conference on Research and development in information retrieval (SIGIR 1998)</title>
				<meeting>the 21st annual international ACM SIGIR conference on Research and development in information retrieval (SIGIR 1998)</meeting>
		<imprint>
			<biblScope unit="page" from="275" to="281" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b119">
	<analytic>
		<title level="a" type="main">A Study of Smoothing Methods for Language Models Applied to Information Retrieval</title>
		<author>
			<persName><forename type="first">C.-X</forename><surname>Zhai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lafferty</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM Transaction of Information System</title>
		<imprint>
			<biblScope unit="page" from="179" to="214" />
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b120">
	<analytic>
		<title level="a" type="main">The Mathematics of Statistical Machine Translation: Parameter Estimation</title>
		<author>
			<persName><forename type="first">P.-F</forename><surname>Brown</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">J D</forename><surname>Pietra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">A D</forename><surname>Pietra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R.-L</forename><surname>Mercer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Computational Linguist</title>
		<imprint>
			<biblScope unit="page" from="263" to="311" />
			<date type="published" when="1993">1993</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b121">
	<analytic>
		<title level="a" type="main">Information Retrieval as Statistical Translation</title>
		<author>
			<persName><forename type="first">A</forename><surname>Berger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lafferty</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 22nd annual international ACM SIGIR conference on Research and development in information retrieval (SIGIR 1999)</title>
				<meeting>the 22nd annual international ACM SIGIR conference on Research and development in information retrieval (SIGIR 1999)</meeting>
		<imprint>
			<biblScope unit="page" from="222" to="229" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b122">
	<analytic>
		<title level="a" type="main">Retrieval Models for Question and Answer Archives</title>
		<author>
			<persName><forename type="first">X</forename><surname>Xue</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Jeon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W.-B</forename><surname>Croft</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 31st annual international ACM SIGIR conference on Research and development in information retrieval (SIGIR 2008)</title>
				<meeting>the 31st annual international ACM SIGIR conference on Research and development in information retrieval (SIGIR 2008)</meeting>
		<imprint>
			<biblScope unit="page" from="475" to="482" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b123">
	<analytic>
		<title level="a" type="main">Mining association rules between sets of items in large databases</title>
		<author>
			<persName><forename type="first">Rakesh</forename><surname>Agrawal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Tomasz</forename><surname>Imieliński</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Arun</forename><surname>Swami</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">SIGMOD Rec</title>
		<imprint>
			<biblScope unit="volume">22</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="207" to="216" />
			<date type="published" when="1993">1993</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b124">
	<monogr>
		<title level="m" type="main">Generalized cores</title>
		<author>
			<persName><forename type="first">V</forename><surname>Batagelj</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Zaveršnik</surname></persName>
		</author>
		<idno>arxiv:cs.DS/0202039</idno>
		<imprint>
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b125">
	<analytic>
		<title level="a" type="main">Evaluating collaborative filtering recommender systems</title>
		<author>
			<persName><forename type="first">Jonathan</forename><forename type="middle">L</forename><surname>Herlocker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Joseph</forename><forename type="middle">A</forename><surname>Konstan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Loren</forename><forename type="middle">G</forename><surname>Terveen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">John</forename><forename type="middle">T</forename><surname>Riedl</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM Trans. Inf. Syst</title>
		<imprint>
			<biblScope unit="volume">22</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="5" to="53" />
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b126">
	<analytic>
		<title level="a" type="main">Bib-Sonomy: A social bookmark and publication sharing system</title>
		<author>
			<persName><forename type="first">Andreas</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Robert</forename><surname>Jäschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Christoph</forename><surname>Schmitz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gerd</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. the First Conceptual Structures Tool Interoperability Workshop at the 14th Int. Conf. on Conceptual Structures</title>
				<meeting>the First Conceptual Structures Tool Interoperability Workshop at the 14th Int. Conf. on Conceptual Structures<address><addrLine>Aalborg</addrLine></address></meeting>
		<imprint>
			<publisher>Aalborg Universitetsforlag</publisher>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="87" to="102" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b127">
	<analytic>
		<title level="a" type="main">Trend detection in folksonomies</title>
		<author>
			<persName><forename type="first">Andreas</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Robert</forename><surname>Jäschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Christoph</forename><surname>Schmitz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gerd</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. First International Conference on Semantics And Digital Media Technology (SAMT)</title>
				<meeting>First International Conference on Semantics And Digital Media Technology (SAMT)</meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2006-12">dec 2006</date>
			<biblScope unit="volume">4306</biblScope>
			<biblScope unit="page" from="56" to="70" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b128">
	<analytic>
		<title level="a" type="main">Tag recommendations in folksonomies</title>
		<author>
			<persName><forename type="first">Robert</forename><surname>Jäschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Leandro</forename><forename type="middle">Balby</forename><surname>Marinho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Andreas</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lars</forename><surname>Schmidt-Thieme</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gerd</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Knowledge Discovery in Databases: PKDD 2007, 11th European Conference on Principles and Practice of Knowledge Discovery in Databases</title>
				<meeting><address><addrLine>Warsaw, Poland</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2007">September 17-21, 2007. 2007</date>
			<biblScope unit="volume">4702</biblScope>
			<biblScope unit="page" from="506" to="514" />
		</imprint>
	</monogr>
	<note>Proceedings</note>
</biblStruct>

<biblStruct xml:id="b129">
	<analytic>
		<title level="a" type="main">Automatic tag recommendation for the web 2.0 blogosphere using collaborative tagging and hybrid ANN semantic structures</title>
		<author>
			<persName><forename type="first">Sigma</forename><surname>On</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kee</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Andy</forename><surname>Hon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Wai</forename><surname>Chun</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ACOS&apos;07: Proc. the 6th Conf. on WSEAS Int. Conf. on Applied Computer Science</title>
				<meeting><address><addrLine>Stevens Point, Wisconsin, USA</addrLine></address></meeting>
		<imprint>
			<publisher>WSEAS</publisher>
			<date type="published" when="2007">2007</date>
			<biblScope unit="page" from="88" to="93" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b130">
	<analytic>
		<title level="a" type="main">Tag recommendation for folksonomies oriented towards individual users</title>
		<author>
			<persName><forename type="first">Marek</forename><surname>Lipczak</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the ECML/PKDD 2008 Discovery Challenge Workshop, part of the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases</title>
				<meeting>the ECML/PKDD 2008 Discovery Challenge Workshop, part of the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b131">
	<analytic>
		<title level="a" type="main">Flickr tag recommendation based on collective knowledge</title>
		<author>
			<persName><forename type="first">Börkur</forename><surname>Sigurbjörnsson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Roelof</forename><surname>Van Zwol</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">WWW &apos;08: Proc. the 17th international conference on World Wide Web</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="327" to="336" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b132">
	<analytic>
		<title level="a" type="main">TagAssist: Automatic tag suggestion for blog posts</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">C</forename><surname>Sood</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">J</forename><surname>Hammond</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">H</forename><surname>Owsley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Birnbaum</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. the International Conference on Weblogs and Social Media (ICWSM</title>
				<meeting>the International Conference on Weblogs and Social Media (ICWSM</meeting>
		<imprint>
			<date type="published" when="2007">2007. 2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b133">
	<analytic>
		<title level="a" type="main">RSDC08: Tag recommendations using bookmark content</title>
		<author>
			<persName><forename type="first">Marta</forename><surname>Tatu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Munirathnam</forename><surname>Srikanth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Thomas</forename><surname>Dsilva</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the ECML/PKDD 2008 Discovery Challenge Workshop, part of the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases</title>
				<meeting>the ECML/PKDD 2008 Discovery Challenge Workshop, part of the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b134">
	<analytic>
		<title level="a" type="main">Contag: A semantic tag recommendation system</title>
		<author>
			<persName><forename type="first">B</forename><surname>Adrian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Sauermann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Roth-Berghofer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of I-Semantics&apos; 07</title>
				<meeting>I-Semantics&apos; 07</meeting>
		<imprint>
			<date type="published" when="2007">2007</date>
			<biblScope unit="page" from="297" to="304" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b135">
	<analytic>
		<title level="a" type="main">Coolrank: A social solution for ranking bookmarked web resources</title>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">S</forename><surname>Al-Khalifa</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Innovations &apos;07. 4th International Conference on</title>
				<imprint>
			<date type="published" when="2007">2007. 2007</date>
			<biblScope unit="page" from="208" to="212" />
		</imprint>
	</monogr>
	<note>Innovations in Information Technology</note>
</biblStruct>

<biblStruct xml:id="b136">
	<analytic>
		<title level="a" type="main">Recommending smart tags in a social bookmarking system</title>
		<author>
			<persName><forename type="first">Pierpaolo</forename><surname>Basile</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Domenico</forename><surname>Gendarmi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Filippo</forename><surname>Lanubile</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Giovanni</forename><surname>Semeraro</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Bridging the Gep between Semantic Web and Web 2</title>
				<meeting><address><addrLine>SemNet</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2007">2007. 2007</date>
			<biblScope unit="page" from="22" to="29" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b137">
	<analytic>
		<title level="a" type="main">Improved annotation of the blogosphere via autotagging and hierarchical clustering</title>
		<author>
			<persName><forename type="first">Christopher</forename><forename type="middle">H</forename><surname>Brooks</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Nancy</forename><surname>Montanez</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">WWW &apos;06: Proceedings of the 15th international conference on World Wide Web</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="625" to="632" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b138">
	<analytic>
		<title level="a" type="main">Trust region newton method for logistic regression</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">S</forename><surname>Keerthi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">J</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">C</forename><surname>Weng</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Machine Learning Research</title>
		<imprint>
			<biblScope unit="issue">9</biblScope>
			<biblScope unit="page" from="627" to="650" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b139">
	<analytic>
		<title level="a" type="main">A kernel method for multi-labelled classification</title>
		<author>
			<persName><forename type="first">Andre</forename><surname>Elisseeff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jason</forename><surname>Weston</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Neural Information Processing Systems 14</title>
				<imprint>
			<publisher>MIT Press</publisher>
			<date type="published" when="2001">2001</date>
			<biblScope unit="page" from="681" to="687" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b140">
	<analytic>
		<title level="a" type="main">Liblinear: A library for large linear classification</title>
		<author>
			<persName><forename type="first">Kai-Wei</forename><surname>Rong-En Fan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Cho-Jui</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Xiang-Rui</forename><surname>Hsieh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Chih-Jen</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><surname>Lin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Machine Learning Research</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="page" from="1871" to="1874" />
			<date type="published" when="2008-08">August 2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b141">
	<analytic>
		<title level="a" type="main">Discriminative methods for multi-labeled classification</title>
		<author>
			<persName><forename type="first">Shantanu</forename><surname>Godbole</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Sunita</forename><surname>Sarawagi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">tors, PAKDD</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">Honghua</forename><surname>Dai</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Ramakrishnan</forename><surname>Srikant</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Chengqi</forename><surname>Zhang</surname></persName>
		</editor>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2004">2004</date>
			<biblScope unit="volume">3056</biblScope>
			<biblScope unit="page" from="22" to="30" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b142">
	<analytic>
		<title level="a" type="main">Trend detection in folksonomies</title>
		<author>
			<persName><forename type="first">Andreas</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Robert</forename><surname>Jäschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Christoph</forename><surname>Schmitz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gerd</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the First International Conference on Semantics And Digital Media Technology (SAMT)</title>
				<editor>
			<persName><forename type="first">S</forename><surname>Yannis</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Yiannis</forename><surname>Avrithis</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Steffen</forename><surname>Kompatsiaris</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Noel</forename><forename type="middle">E</forename><surname>Staab</surname></persName>
		</editor>
		<editor>
			<persName><surname>O'connor</surname></persName>
		</editor>
		<meeting>the First International Conference on Semantics And Digital Media Technology (SAMT)<address><addrLine>Heidelberg</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2006">12 2006</date>
			<biblScope unit="volume">4306</biblScope>
			<biblScope unit="page" from="56" to="70" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b143">
	<analytic>
		<title level="a" type="main">Tag recommendations in folksonomies</title>
		<author>
			<persName><forename type="first">Robert</forename><surname>Jäschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Leandro</forename><surname>Marinho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Andreas</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lars</forename><surname>Schmidt-Thieme</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gerd</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Workshop Proceedings of Lernen -Wissensentdeckung -Adaptivit (LWA 2007)</title>
				<editor>
			<persName><forename type="first">Alexander</forename><surname>Hinneburg</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="2007-09">sep 2007</date>
			<biblScope unit="page" from="13" to="20" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b144">
	<analytic>
		<title level="a" type="main">Optimizing search engines using clickthrough data</title>
		<author>
			<persName><forename type="first">Thorsten</forename><surname>Joachims</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">KDD &apos;02: Proceedings of the eighth ACM SIGKDD international conference on Knowledge discovery and data mining</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2002">2002</date>
			<biblScope unit="page" from="133" to="142" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b145">
	<analytic>
		<title level="a" type="main">Multilabel text classification for automated tag suggestion</title>
		<author>
			<persName><forename type="first">Ioannis</forename><surname>Katakis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Grigorios</forename><surname>Tsoumakas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ioannis</forename><surname>Vlahavas</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of ECML PKDD Discovery Challenge (RSDC08)</title>
				<meeting>ECML PKDD Discovery Challenge (RSDC08)</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="75" to="83" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b146">
	<monogr>
		<title level="m" type="main">The art of tagging: Measuring the quality of tags</title>
		<author>
			<persName><forename type="first">Ralf</forename><surname>Krestel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ling</forename><surname>Chen</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="257" to="271" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b147">
	<analytic>
		<title level="a" type="main">Automatic tag recommendation for the web 2.0 blogosphere using collaborative tagging and hybrid ann semantic structures</title>
		<author>
			<persName><forename type="first">Sigma</forename><surname>On</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kee</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Andy</forename><surname>Hon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Wai</forename><surname>Chun</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 6th Conference on WSEAS International Conference on Applied Computer Science</title>
		<title level="s">World Scientific and Engineering Academy and Society</title>
		<meeting>the 6th Conference on WSEAS International Conference on Applied Computer Science<address><addrLine>Stevens Point, Wisconsin, USA</addrLine></address></meeting>
		<imprint>
			<publisher>WSEAS</publisher>
			<date type="published" when="2007">2007</date>
			<biblScope unit="page" from="88" to="93" />
		</imprint>
	</monogr>
	<note>ACOS&apos;07</note>
</biblStruct>

<biblStruct xml:id="b148">
	<analytic>
		<title level="a" type="main">Collaborative tag recommendations</title>
		<author>
			<persName><forename type="first">Leandro</forename><surname>Balby</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Marinho</forename></persName>
		</author>
		<author>
			<persName><forename type="first">Lars</forename><surname>Schmidt-Thieme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Studies in Classification, Data Analysis, and Knowledge Organization</title>
				<meeting><address><addrLine>Berlin Heidelberg</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="533" to="540" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b149">
	<analytic>
		<title level="a" type="main">Autotag: a collaborative approach to automated tag assignment for weblog posts</title>
		<author>
			<persName><forename type="first">Gilad</forename><surname>Mishne</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 15th International Conference on World Wide Web</title>
				<meeting>the 15th International Conference on World Wide Web<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM Press</publisher>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="953" to="954" />
		</imprint>
	</monogr>
	<note>WWW &apos;06. paper presented at the poster track</note>
</biblStruct>

<biblStruct xml:id="b150">
	<analytic>
		<title level="a" type="main">Feature selection for unbalanced class distribution and naive bayes</title>
		<author>
			<persName><forename type="first">D</forename><surname>Mladenic</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Grobelnik</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. 16th International Conference on Machine Learning ICML-99</title>
				<meeting>16th International Conference on Machine Learning ICML-99<address><addrLine>Bled, SL</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1999">1999</date>
			<biblScope unit="page" from="258" to="267" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b151">
	<analytic>
		<title level="a" type="main">Scoring and selecting terms for text categorization</title>
		<author>
			<persName><forename type="first">E</forename><surname>Montañés</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Díaz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ranilla</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">F</forename><surname>Combarro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Fernández</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Intelligent Systems</title>
		<imprint>
			<biblScope unit="volume">20</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="40" to="47" />
			<date type="published" when="2005-06">May/June 2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b152">
	<analytic>
		<title level="a" type="main">Probabilistic outputs for support vector machines and comparisons to regularized likelihood methods</title>
		<author>
			<persName><forename type="first">John</forename><forename type="middle">C</forename><surname>Platt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">John</forename><forename type="middle">C</forename><surname>Platt</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Large Margin Classifiers</title>
				<imprint>
			<publisher>MIT Press</publisher>
			<date type="published" when="1999">1999</date>
			<biblScope unit="page" from="61" to="74" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b153">
	<analytic>
		<title level="a" type="main">Fan: Finding accurate inductions</title>
		<author>
			<persName><forename type="first">J</forename><surname>Ranilla</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bahamonde</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Human Computer Studies</title>
		<imprint>
			<biblScope unit="volume">56</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="445" to="474" />
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b154">
	<monogr>
		<title level="m" type="main">An introduction to modern information retrieval</title>
		<author>
			<persName><forename type="first">G</forename><surname>Salton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">J</forename><surname>Mcgill</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1983">1983</date>
			<publisher>McGraw-Hill</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b155">
	<analytic>
		<title level="a" type="main">Machine learning in automated text categorisation</title>
		<author>
			<persName><forename type="first">F</forename><surname>Sebastiani</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM Computing Survey</title>
		<imprint>
			<biblScope unit="volume">34</biblScope>
			<biblScope unit="issue">1</biblScope>
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b156">
	<analytic>
		<title level="a" type="main">Flickr tag recommendation based on collective knowledge</title>
		<author>
			<persName><forename type="first">Börkur</forename><surname>Sigurbjörnsson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Roelof</forename><surname>Van Zwol</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">WWW &apos;08: Proceeding of the 17th international conference on World Wide Web</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="327" to="336" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b157">
	<analytic>
		<title level="a" type="main">Tagassist: Automatic tag suggestion for blog posts</title>
		<author>
			<persName><forename type="first">Sanjay</forename><surname>Sood</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Sara</forename><surname>Owsley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kristian</forename><surname>Hammond</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Larry</forename><surname>Birnbaum</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the International Conference on Weblogs and Social Media (ICWSM 2007)</title>
				<meeting>the International Conference on Weblogs and Social Media (ICWSM 2007)</meeting>
		<imprint>
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b158">
	<analytic>
		<title level="a" type="main">Rsdc&apos;08: Tag recommendations using bookmark content</title>
		<author>
			<persName><forename type="first">M</forename><surname>Tatu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Srikanth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">D</forename><surname>Silva</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of ECML PKDD Discovery Challenge (RSDC08)</title>
				<meeting>ECML PKDD Discovery Challenge (RSDC08)</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="96" to="107" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b159">
	<analytic>
		<title level="a" type="main">Multi label classification: An overview</title>
		<author>
			<persName><forename type="first">G</forename><surname>Tsoumakas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Katakis</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Data Warehousing and Mining</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="1" to="13" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b160">
	<monogr>
		<title level="m" type="main">Ranking and suggesting tags in collaborative tagging applications</title>
		<author>
			<persName><forename type="first">M</forename><surname>Vojnovic</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Cruise</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Gunawardena</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Marbach</surname></persName>
		</author>
		<idno>MSR-TR-2007-06</idno>
		<imprint>
			<date type="published" when="2007">2007</date>
		</imprint>
		<respStmt>
			<orgName>Microsoft Research</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Technical Report</note>
</biblStruct>

<biblStruct xml:id="b161">
	<analytic>
		<title level="a" type="main">Towards the semantic web: Collaborative tag suggestions</title>
		<author>
			<persName><forename type="first">Zhichen</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yun</forename><surname>Fu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jianchang</forename><surname>Mao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Difu</forename><surname>Su</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">WWW2006: Proceedings of the Collaborative Web Tagging Workshop</title>
				<meeting><address><addrLine>Edinburgh, Scotland</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b162">
	<analytic>
		<title level="a" type="main">Improved Recommendation based on Collaborative Tagging Behaviors</title>
		<author>
			<persName><forename type="first">Shiwan</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Nan</forename><surname>Du</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Andreas</forename><surname>Nauerz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Xiatian</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Quan</forename><surname>Yuan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Rongyao</forename><surname>Fu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the International ACM Conference on Intelligent User Interfaces (IUI2008)</title>
				<meeting>the International ACM Conference on Intelligent User Interfaces (IUI2008)<address><addrLine>Canary Islands, Spain</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b163">
	<analytic>
		<title level="a" type="main">Tag Recommendations in Folksonomies</title>
		<author>
			<persName><forename type="first">R</forename><surname>Jäschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">B</forename><surname>Marinho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Schmidt-Thieme</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">PKDD 2007</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b164">
	<analytic>
		<title level="a" type="main">Learning to Recognize Valuable Tags</title>
		<author>
			<persName><forename type="first">S</forename><surname>Sen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Vig</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Rield</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IUT&apos;09</title>
				<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b165">
	<analytic>
		<title level="a" type="main">Enhancing Information Scent: Identifying and Recommending Quality Tags</title>
		<author>
			<persName><forename type="first">S</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Farooq</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Carroll</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">GROUP&apos;09</title>
				<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b166">
	<analytic>
		<title level="a" type="main">Using noun phrase heads to extract document keyphrases</title>
		<author>
			<persName><forename type="first">K</forename><surname>Barker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Cornacchia</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 13th Canadian Conference on Artificial Intelligence</title>
				<meeting>the 13th Canadian Conference on Artificial Intelligence<address><addrLine>Montreal, Canada</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2000">2000</date>
			<biblScope unit="page" from="417" to="426" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b167">
	<analytic>
		<title level="a" type="main">Algorithms and applications for approximate nonnegative matrix factorization</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">W</forename><surname>Berry</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Browne</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">N</forename><surname>Langville</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Pauca</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">J</forename><surname>Plemmons</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computational Statistics and Data Analysis</title>
		<imprint>
			<biblScope unit="volume">15</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="155" to="173" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b168">
	<analytic>
		<title level="a" type="main">Using linear algebra for intelligent information retrieval</title>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">M</forename><surname>Berry</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">T</forename><surname>Dumais</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">W</forename><surname>O'brien</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">SIAM Review</title>
		<imprint>
			<biblScope unit="volume">37</biblScope>
			<biblScope unit="page" from="573" to="595" />
			<date type="published" when="1995">1995</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b169">
	<analytic>
		<title level="a" type="main">Svd-based initialization: A head start on nonnegative matrix factorization</title>
		<author>
			<persName><forename type="first">C</forename><surname>Boutsidis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Gallopoulos</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Pattern Recognition</title>
		<imprint>
			<biblScope unit="volume">41</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="1350" to="1362" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b170">
	<analytic>
		<title level="a" type="main">Using concept lattices for text retrieval and mining</title>
		<author>
			<persName><forename type="first">C</forename><surname>Carpineto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Romano</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Formal Concept Analysis</title>
				<editor>
			<persName><forename type="first">B</forename><forename type="middle">G</forename></persName>
		</editor>
		<imprint>
			<publisher>Springer-Verlag</publisher>
			<date type="published" when="2005">2005</date>
			<biblScope unit="page" from="161" to="179" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b171">
	<monogr>
		<title level="m" type="main">Optimality, computation, and interpretations of nonnegative matrix factorization</title>
		<author>
			<persName><forename type="first">M</forename><surname>Chu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Diele</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Plemmons</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ragni</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2005">2005</date>
		</imprint>
		<respStmt>
			<orgName>Department of Mathematics, North Carolina State University</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Technical report</note>
</biblStruct>

<biblStruct xml:id="b172">
	<analytic>
		<title level="a" type="main">Indexing by latent semantic analysis</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">C</forename><surname>Deerwester</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">T</forename><surname>Dumais</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">K</forename><surname>Landauer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">W</forename><surname>Furnas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">A</forename><surname>Harshman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of the American Society of Information Science</title>
		<imprint>
			<biblScope unit="volume">41</biblScope>
			<biblScope unit="issue">6</biblScope>
			<biblScope unit="page" from="391" to="407" />
			<date type="published" when="1990">1990</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b173">
	<analytic>
		<title level="a" type="main">Domain-specific keyphrase extraction</title>
		<author>
			<persName><forename type="first">E</forename><surname>Frank</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">W</forename><surname>Paynter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">H</forename><surname>Witten</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Gutwin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">G</forename><surname>Nevill-Manning</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the International Joint Conference on Artificial Intelligence</title>
				<meeting>the International Joint Conference on Artificial Intelligence</meeting>
		<imprint>
			<date type="published" when="1999">1999</date>
			<biblScope unit="page" from="668" to="673" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b174">
	<analytic>
		<title level="a" type="main">Usage patterns of collaborative tagging systems</title>
		<author>
			<persName><forename type="first">S</forename><surname>Golder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Huberman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Information Science</title>
		<imprint>
			<biblScope unit="volume">32</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="198" to="208" />
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b175">
	<analytic>
		<title level="a" type="main">Social tag prediction</title>
		<author>
			<persName><forename type="first">P</forename><surname>Heymann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Ramage</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">G</forename><surname>Molina</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceeding of SIGIR</title>
				<meeting>eeding of SIGIR</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b176">
	<analytic>
		<title level="a" type="main">Information retrieval in folksonomies: Search and ranking</title>
		<author>
			<persName><forename type="first">A</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Jaschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Schmitz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceeding of the 3rd European Semantic Web Conference</title>
				<meeting>eeding of the 3rd European Semantic Web Conference</meeting>
		<imprint>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="411" to="426" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b177">
	<analytic>
		<title level="a" type="main">Non-negative matrix factorization with sparseness constraints</title>
		<author>
			<persName><forename type="first">P</forename><surname>Hoyer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Machine Learning Research</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="page" from="1457" to="1469" />
			<date type="published" when="1994">1994</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b178">
	<monogr>
		<title level="m" type="main">Combining Machine Learning and Natural Language Processing for Automatic Keyword Extraction</title>
		<author>
			<persName><forename type="first">A</forename><surname>Hulth</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2004">2004</date>
		</imprint>
		<respStmt>
			<orgName>Stockholm University</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">PhD thesis</note>
</biblStruct>

<biblStruct xml:id="b179">
	<analytic>
		<title level="a" type="main">Tag recommendations in folksonomies</title>
		<author>
			<persName><forename type="first">R</forename><surname>Jaschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Marinho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">S</forename><surname>Thieme</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">The 11th European Conference on Principles and Practice of Knowledge Discovery in Databases</title>
				<editor>
			<persName><forename type="first">J</forename><forename type="middle">N</forename><surname>Kok</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Koronacki</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><forename type="middle">L</forename><surname>De Ma'ntaras</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Matwin</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Mladenic</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Skowron</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="2007">2007</date>
			<biblScope unit="page" from="506" to="514" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b180">
	<analytic>
		<title level="a" type="main">Multilabel text classification for automated tag suggestion</title>
		<author>
			<persName><forename type="first">I</forename><surname>Katakis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Tsoumakas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Vlahavas</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceeding of the ECML/PKDD Discovery Challenge</title>
				<meeting>eeding of the ECML/PKDD Discovery Challenge</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b181">
	<monogr>
		<title level="m" type="main">Indexing and Abstracting in Theory and Practice</title>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">W</forename><surname>Lancaster</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1998">1998</date>
			<publisher>Library Association Publishing</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b182">
	<analytic>
		<title level="a" type="main">Learning the parts of objects by nonnegative matrix factorization</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">D</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">S</forename><surname>Seung</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Nature</title>
		<imprint>
			<biblScope unit="volume">401</biblScope>
			<biblScope unit="page" from="788" to="791" />
			<date type="published" when="1999">1999</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b183">
	<analytic>
		<title level="a" type="main">Algorithms for non-negative matrix factorization</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">D</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">S</forename><surname>Seung</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Neural Information Processing System</title>
				<imprint>
			<date type="published" when="2001">2001</date>
			<biblScope unit="page" from="556" to="562" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b184">
	<analytic>
		<title level="a" type="main">On the convergence of multiplicative update algorithms for non-negative matrix factorization</title>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">J</forename><surname>Lin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Neural Networks</title>
		<imprint>
			<biblScope unit="volume">18</biblScope>
			<biblScope unit="page" from="1589" to="1596" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b185">
	<analytic>
		<title level="a" type="main">Projected gradient methods for non-negative matrix factorization</title>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">J</forename><surname>Lin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Neural Computation</title>
		<imprint>
			<biblScope unit="volume">19</biblScope>
			<biblScope unit="page" from="2756" to="2779" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b186">
	<analytic>
		<title level="a" type="main">Concept discovery from text</title>
		<author>
			<persName><forename type="first">D</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Pantel</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the International Conference on Computational Linguistics</title>
				<meeting>the International Conference on Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2002">2002</date>
			<biblScope unit="page" from="577" to="583" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b187">
	<analytic>
		<title level="a" type="main">Domain-independent automatic keyphrase indexing with small training set</title>
		<author>
			<persName><forename type="first">O</forename><surname>Medelyan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">H</forename><surname>Witten</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of the American Society for Information Science and Technology</title>
		<imprint>
			<biblScope unit="volume">59</biblScope>
			<biblScope unit="issue">7</biblScope>
			<biblScope unit="page" from="1026" to="1040" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b188">
	<analytic>
		<title level="a" type="main">Autotag: A collaborative approach to automated tag assignment for weblog posts</title>
		<author>
			<persName><forename type="first">G</forename><surname>Mishne</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceeding of the 15th International Conference on World Wide Web</title>
				<meeting>eeding of the 15th International Conference on World Wide Web</meeting>
		<imprint>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="953" to="954" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b189">
	<analytic>
		<title level="a" type="main">Positive matrix factorization: A non-negative factor model with optimal utilization of error estimates of data values</title>
		<author>
			<persName><forename type="first">P</forename><surname>Paatero</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Tapper</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Environmetrics</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="page" from="111" to="126" />
			<date type="published" when="1994">1994</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b190">
	<analytic>
		<title level="a" type="main">A three-pronged approach to the extraction of key terms and semantic roles</title>
		<author>
			<persName><forename type="first">C</forename><surname>Paice</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Black</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the International Conference on Recent Advances in NLP</title>
				<meeting>the International Conference on Recent Advances in NLP<address><addrLine>Borovets, Bulgaria</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2003">2003</date>
			<biblScope unit="page" from="357" to="363" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b191">
	<analytic>
		<title level="a" type="main">Formal concept analysis in information science</title>
		<author>
			<persName><forename type="first">U</forename><surname>Priss</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Annual Review of Information Science and Technology</title>
		<imprint>
			<biblScope unit="volume">40</biblScope>
			<biblScope unit="page" from="521" to="543" />
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b192">
	<analytic>
		<title level="a" type="main">Unsupervised text learning for the learning of dogma-inspired ontologies</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">L</forename><surname>Reinberger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Spyns</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Ontology Learning from Text: Method, Application and Evaluation</title>
				<editor>
			<persName><forename type="first">P</forename><surname>Buitelaar</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Cimiano</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Magnini</surname></persName>
		</editor>
		<imprint>
			<publisher>IOS Press</publisher>
			<date type="published" when="2005">2005</date>
			<biblScope unit="page" from="29" to="43" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b193">
	<analytic>
		<title level="a" type="main">Tagassist: Automatic tag suggestion for blog posts</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">C</forename><surname>Sood</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">H</forename><surname>Owsley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">J</forename><surname>Hammond</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Birnbaum</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceeding of the International Conference on Weblogs and Social Media (ICWSM)</title>
				<meeting>eeding of the International Conference on Weblogs and Social Media (ICWSM)</meeting>
		<imprint>
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b194">
	<analytic>
		<title level="a" type="main">Rsdc&apos;08: Tag recommendation using bookmark content</title>
		<author>
			<persName><forename type="first">M</forename><surname>Tatu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Srikanth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">D</forename><surname>Silva</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceeding of the ECML/PKDD Discovery Challenge</title>
				<meeting>eeding of the ECML/PKDD Discovery Challenge</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b195">
	<analytic>
		<title level="a" type="main">Learning algorithms for keyphrase extraction</title>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">D</forename><surname>Turney</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Information Retrieval</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="303" to="336" />
			<date type="published" when="2000">2000</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b196">
	<analytic>
		<title level="a" type="main">Towards the semantic web: Collaborative tag suggestions</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Fu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Mao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Su</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceeding of Collaborative Web Tagging Workshop</title>
				<meeting>eeding of Collaborative Web Tagging Workshop</meeting>
		<imprint>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b197">
	<monogr>
		<title level="m" type="main">Modern Information Retrieval</title>
		<author>
			<persName><forename type="first">R</forename><surname>Baeza-Yates</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Ribeiro-Neto</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1999">1999</date>
			<publisher>Addison-Wesley</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b198">
	<analytic>
		<title level="a" type="main">Learning collaborative information filters</title>
		<author>
			<persName><forename type="first">D</forename><surname>Billsus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">J</forename><surname>Pazzani</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceeding of the 15th International Conference on Machine Learning</title>
				<meeting>eeding of the 15th International Conference on Machine Learning</meeting>
		<imprint>
			<biblScope unit="page" from="46" to="54" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b199">
	<monogr>
		<title/>
		<author>
			<persName><forename type="first">Morgan</forename><surname>Kaufmann</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1998">1998</date>
			<pubPlace>San Francisco, CA</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b200">
	<analytic>
		<title level="a" type="main">Improved annotation of the blogosphere via autotagging and hierarchical clustering</title>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">H</forename><surname>Brooks</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Montanez</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 15th international conference on World Wide Web</title>
				<meeting>the 15th international conference on World Wide Web<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM Press</publisher>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="625" to="632" />
		</imprint>
	</monogr>
	<note>WWW &apos;06</note>
</biblStruct>

<biblStruct xml:id="b201">
	<analytic>
		<title level="a" type="main">Network properties of folksonomies</title>
		<author>
			<persName><forename type="first">C</forename><surname>Cattuto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Schmitz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Baldassarri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">D P</forename><surname>Servedio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Loreto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Grahl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">AI Communications</title>
		<imprint>
			<biblScope unit="volume">20</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="245" to="262" />
			<date type="published" when="2007-12">December 2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b202">
	<analytic>
		<title level="a" type="main">The Structure of Collaborative Tagging Systems</title>
		<author>
			<persName><forename type="first">S</forename><surname>Golder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">A</forename><surname>Huberman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Information Science</title>
		<imprint>
			<biblScope unit="volume">32</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="198" to="208" />
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b203">
	<analytic>
		<title level="a" type="main">Social tag prediction</title>
		<author>
			<persName><forename type="first">P</forename><surname>Heymann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Ramage</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Garcia-Molina</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">SIGIR &apos;08: Proceedings of the 31st annual international ACM SIGIR conference on Research and development in information retrieval</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="531" to="538" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b204">
	<analytic>
		<title level="a" type="main">Tag recommendations in folksonomies</title>
		<author>
			<persName><forename type="first">R</forename><surname>Jäschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Marinho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Schmidt-Thieme</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Workshop Proceedings of Lernen -Wissensentdeckung -Adaptivit?t (LWA 2007)</title>
				<editor>
			<persName><forename type="first">Alexander</forename><surname>Hinneburg</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="2007-09">September 2007</date>
			<biblScope unit="page" from="13" to="20" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b205">
	<analytic>
		<title level="a" type="main">Automatic tag recommendation for the web 2.0 blogosphere using collaborative tagging and hybrid ann semantic structures</title>
		<author>
			<persName><forename type="first">Sigma</forename><surname>On</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kee</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Andy</forename><surname>Hon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Wai</forename><surname>Chun</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 6th Conference on WSEAS International Conference on Applied Computer Science</title>
		<title level="s">World Scientific and Engineering Academy and Society</title>
		<meeting>the 6th Conference on WSEAS International Conference on Applied Computer Science<address><addrLine>Stevens Point, Wisconsin, USA</addrLine></address></meeting>
		<imprint>
			<publisher>WSEAS</publisher>
			<date type="published" when="2007">2007</date>
			<biblScope unit="page" from="88" to="93" />
		</imprint>
	</monogr>
	<note>ACOS&apos;07</note>
</biblStruct>

<biblStruct xml:id="b206">
	<analytic>
		<title level="a" type="main">Tag recommendation for folksonomies oriented towards individual users</title>
		<author>
			<persName><forename type="first">M</forename><surname>Lipczak</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of ECML PKDD Discovery Challenge (RSDC08)</title>
				<meeting>ECML PKDD Discovery Challenge (RSDC08)</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="84" to="95" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b207">
	<monogr>
		<title level="m" type="main">Collaborative tag recommendations</title>
		<author>
			<persName><forename type="first">Leandro</forename><forename type="middle">B</forename><surname>Marinho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lars</forename><surname>Schmidt-Thieme</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="533" to="540" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b208">
	<monogr>
		<author>
			<persName><forename type="first">Adam</forename><surname>Mathes</surname></persName>
		</author>
		<ptr target="http://www.adammathes.com/academic/computer-mediated-communication/folksonomies.html" />
		<title level="m">Folksonomies -cooperative classification and communication through shared metadata</title>
				<imprint>
			<date type="published" when="2004-12">December 2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b209">
	<analytic>
		<title level="a" type="main">Autotag: a collaborative approach to automated tag assignment for weblog posts</title>
		<author>
			<persName><forename type="first">Gilad</forename><surname>Mishne</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 15th international conference on World Wide Web</title>
				<meeting>the 15th international conference on World Wide Web<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="953" to="954" />
		</imprint>
	</monogr>
	<note>WWW &apos;06</note>
</biblStruct>

<biblStruct xml:id="b210">
	<analytic>
		<title level="a" type="main">Okapi at trec</title>
		<author>
			<persName><forename type="first">Stephen</forename><forename type="middle">E</forename><surname>Robertson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Steve</forename><surname>Walker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Micheline</forename><forename type="middle">H</forename><surname>Beaulieu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Aarron</forename><surname>Gull</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Marianna</forename><surname>Lau</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Text REtrieval Conference</title>
				<imprint>
			<date type="published" when="1992">1992</date>
			<biblScope unit="page" from="21" to="30" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b211">
	<monogr>
		<title level="m" type="main">Automatic Text Processing</title>
		<author>
			<persName><forename type="first">G</forename><surname>Salton</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1989">1989</date>
			<publisher>Addison-Wesley</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b212">
	<analytic>
		<title level="a" type="main">Mining association rules in folksonomies</title>
		<author>
			<persName><forename type="first">Christoph</forename><surname>Schmitz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Andreas</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Robert</forename><surname>Jschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gerd</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Data Science and Classification (Proc. IFCS 2006 Conference), Studies in Classification, Data Analysis, and Knowledge Organization</title>
				<editor>
			<persName><forename type="first">V</forename><surname>Batagelj</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H.-H</forename><surname>Bock</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Ferligoj</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Iberna</surname></persName>
		</editor>
		<meeting><address><addrLine>Berlin/Heidelberg; Ljubljana</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2006-07">July 2006</date>
			<biblScope unit="page" from="261" to="270" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b213">
	<analytic>
		<title level="a" type="main">TagAssist: Automatic Tag Suggestion for Blog Posts</title>
		<author>
			<persName><forename type="first">Sanjay</forename><surname>Sood</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Sara</forename><surname>Owsley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kristian</forename><surname>Hammond</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Larry</forename><surname>Birnbaum</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the International Conference on Weblogs and Social Media (ICWSM</title>
				<meeting>the International Conference on Weblogs and Social Media (ICWSM</meeting>
		<imprint>
			<date type="published" when="2007">2007. 2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b214">
	<analytic>
		<title level="a" type="main">Rsdc&apos;08: Tag recommendations using bookmark content</title>
		<author>
			<persName><forename type="first">M</forename><surname>Tatu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Srikanth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">D</forename><surname>Silva</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of ECML PKDD Discovery Challenge (RSDC08)</title>
				<meeting>ECML PKDD Discovery Challenge (RSDC08)</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="96" to="107" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b215">
	<monogr>
		<title level="m" type="main">Folksonomy coinage and definition</title>
		<author>
			<persName><forename type="first">Thomas</forename><surname>Vander</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Wal</forename></persName>
		</author>
		<ptr target="http://vanderwal.net/folksonomy.html" />
		<imprint>
			<date type="published" when="2007-02">Februar 2007</date>
			<publisher>Website</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b216">
	<analytic>
		<title level="a" type="main">Harvesting social knowledge from folksonomies</title>
		<author>
			<persName><forename type="first">Harris</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Mohammad</forename><surname>Zubair</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kurt</forename><surname>Maly</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">HYPERTEXT &apos;06: Proceedings of the seventeenth conference on Hypertext and hypermedia</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM Press</publisher>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="111" to="114" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b217">
	<analytic>
		<title level="a" type="main">Tag recommendations in folksonomies</title>
		<author>
			<persName><forename type="first">R</forename><surname>Jäschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">B</forename><surname>Marinho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Schmidt-Thieme</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">PKDD</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">J</forename><forename type="middle">N</forename><surname>Kok</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Koronacki</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><forename type="middle">L</forename><surname>De Mántaras</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Matwin</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Mladenic</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Skowron</surname></persName>
		</editor>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2007">2007</date>
			<biblScope unit="volume">4702</biblScope>
			<biblScope unit="page" from="506" to="514" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b218">
	<analytic>
		<title level="a" type="main">BibSonomy: A social bookmark and publication sharing system</title>
		<author>
			<persName><forename type="first">A</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Jäschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Schmitz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. Workshop on Conceptual Structure Tool Interoperability at the Int. Conf. on Conceptual Structures</title>
				<meeting>Workshop on Conceptual Structure Tool Interoperability at the Int. Conf. on Conceptual Structures</meeting>
		<imprint>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="87" to="102" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b219">
	<monogr>
		<title level="m" type="main">Generalized cores</title>
		<author>
			<persName><forename type="first">V</forename><surname>Batagelj</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Zaversnik</surname></persName>
		</author>
		<idno>cs.DS/0202039</idno>
		<ptr target="http://arxiv.org/abs/cs/0202039" />
		<imprint>
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b220">
	<monogr>
		<author>
			<persName><forename type="first">R</forename><surname>Jaeschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Marinho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Schmidt-Thieme</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Stumme</surname></persName>
		</author>
		<title level="m">Tag recommendations in social bookmarking systems</title>
				<imprint>
			<publisher>AICOM</publisher>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b221">
	<analytic>
		<title level="a" type="main">Learning optimal ranking with tensor factorization for tag recommendation</title>
		<author>
			<persName><forename type="first">S</forename><surname>Rendle</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">B</forename><surname>Marinho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Nanopoulos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">S</forename><surname>Thieme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">KDD &apos;09: Proceeding of the 15th ACM SIGKDD international conference on Knowledge discovery and data mining</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b222">
	<analytic>
		<title level="a" type="main">Bpr: Bayesian personalized ranking from implicit feedback</title>
		<author>
			<persName><forename type="first">S</forename><surname>Rendle</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Freudenthaler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Gantner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">S</forename><surname>Thieme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 25th Conference on Uncertainty in Artificial Intelligence (UAI</title>
				<meeting>the 25th Conference on Uncertainty in Artificial Intelligence (UAI</meeting>
		<imprint>
			<date type="published" when="2009">2009. 2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b223">
	<monogr>
		<title level="m" type="main">The pagerank citation ranking: Bringing order to the web</title>
		<author>
			<persName><forename type="first">L</forename><surname>Page</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Brin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Motwani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Winograd</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1998">1998</date>
		</imprint>
		<respStmt>
			<orgName>Stanford Digital Library Technologies Project</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Technical report</note>
</biblStruct>

<biblStruct xml:id="b224">
	<analytic>
		<title level="a" type="main">Tag recommendations in folksonomies</title>
		<author>
			<persName><forename type="first">R</forename><surname>Jaschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Marinho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Schmidt-Thieme</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 11th European conference on Principles and Practice of Knowledge Discovery in Databases</title>
				<meeting>the 11th European conference on Principles and Practice of Knowledge Discovery in Databases<address><addrLine>Berlin; Heidelberg</addrLine></address></meeting>
		<imprint>
			<publisher>Springer-Verlag</publisher>
			<date type="published" when="2007">2007</date>
			<biblScope unit="page" from="506" to="514" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b225">
	<analytic>
		<title level="a" type="main">Diffusionrank: a possible penicillin for web spamming</title>
		<author>
			<persName><forename type="first">H</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>King</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">R</forename><surname>Lyu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of SIGIR</title>
				<meeting>SIGIR</meeting>
		<imprint>
			<date type="published" when="2007">2007</date>
			<biblScope unit="page" from="431" to="438" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b226">
	<analytic>
		<title level="a" type="main">Mining social networks using heat diffusion processes for marketing candidates selection</title>
		<author>
			<persName><forename type="first">H</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">R</forename><surname>Lyu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>King</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceeding of CIKM</title>
				<meeting>eeding of CIKM</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="233" to="242" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b227">
	<analytic>
		<title level="a" type="main">Learning latent semantic relations from clickthrough data for query suggestion</title>
		<author>
			<persName><forename type="first">H</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>King</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">R</forename><surname>Lyu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceeding of CIKM</title>
				<meeting>eeding of CIKM</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="709" to="718" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b228">
	<monogr>
		<title level="m" type="main">Introduction to information retrieval</title>
		<author>
			<persName><forename type="first">C</forename><surname>Manning</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Raghavan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Schtze</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2008">2008</date>
			<publisher>Cambridge University Press</publisher>
			<pubPlace>New York, NY, USA</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b229">
	<analytic>
		<title level="a" type="main">A comparative study on feature selection in text categorization</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Pedersen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">MACHINE LEARNING-INTERNATIONAL WORKSHOP THEN CONFERENCE</title>
				<imprint>
			<publisher>MORGAN KAUFMANN PUBLISHERS, INC</publisher>
			<date type="published" when="1997">1997</date>
			<biblScope unit="page" from="412" to="420" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b230">
	<analytic>
		<title level="a" type="main">Tag recommendations in social bookmarking systems</title>
		<author>
			<persName><forename type="first">R</forename><surname>Jaschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Marinho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Schmidt-Thieme</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">AI Communications</title>
		<imprint>
			<biblScope unit="volume">21</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="231" to="247" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b231">
	<monogr>
		<title level="m" type="main">Generalized cores</title>
		<author>
			<persName><forename type="first">V</forename><surname>Batagelj</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Zaversnik</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b232">
	<analytic>
		<title level="a" type="main">Autotag: a collaborative approach to automated tag assignment for weblog posts</title>
		<author>
			<persName><forename type="first">G</forename><surname>Mishne</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 15th international conference on World Wide Web</title>
				<meeting>the 15th international conference on World Wide Web<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="953" to="954" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b233">
	<analytic>
		<title level="a" type="main">Browsing system for weblog articles based on automated folksonomy</title>
		<author>
			<persName><forename type="first">T</forename><surname>Ohkura</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Kiyota</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Nakagawa</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the WWW 2006 Workshop on the Weblogging Ecosystem: Aggregation, Analysis and Dynamics, at WWW</title>
				<meeting>the WWW 2006 Workshop on the Weblogging Ecosystem: Aggregation, Analysis and Dynamics, at WWW</meeting>
		<imprint>
			<date type="published" when="2006">2006. 2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b234">
	<analytic>
		<title level="a" type="main">Multilabel text classification for automated tag suggestion</title>
		<author>
			<persName><forename type="first">I</forename><surname>Katakis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Tsoumakas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Vlahavas</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ECML PKDD Discovery Challenge</title>
				<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="volume">75</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b235">
	<analytic>
		<title level="a" type="main">Tag Recommendation for Folksonomies Oriented towards Individual Users</title>
		<author>
			<persName><forename type="first">M</forename><surname>Lipczak</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ECML PKDD Discovery Challenge</title>
				<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="volume">84</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b236">
	<analytic>
		<title level="a" type="main">Rsdc&apos;08: Tag recommendations using bookmark content</title>
		<author>
			<persName><forename type="first">M</forename><surname>Tatu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Srikanth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>D'silva</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ECML PKDD Discovery Challenge</title>
				<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b237">
	<analytic>
		<title level="a" type="main">Towards the semantic web: Collaborative tag suggestions</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Fu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Mao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Su</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">RSDC&apos;09: Tag Recommendation Using Keywords and Association Rules Jian Wang</title>
				<editor>
			<persName><forename type="first">Liangjie</forename><surname>Hong</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Brian</forename><forename type="middle">D</forename><surname>Davison</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
	<note>Collaborative Web Tagging Workshop at WWW2006</note>
</biblStruct>

<biblStruct xml:id="b238">
	<analytic>
		<title level="a" type="main">Social tag prediction</title>
		<author>
			<persName><forename type="first">P</forename><surname>Heymann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Ramage</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Garcia-Molina</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">SIGIR &apos;08: Proceedings of the 31st annual international ACM SIGIR conference on Research and development in information retrieval</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="531" to="538" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b239">
	<analytic>
		<title level="a" type="main">Multilabel text classification for automated tag suggestion</title>
		<author>
			<persName><forename type="first">I</forename><surname>Katakis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Tsoumakas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Vlahavas</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the ECML/PKDD 2008 Discovery Challenge Workshop, part of the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases</title>
				<meeting>the ECML/PKDD 2008 Discovery Challenge Workshop, part of the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b240">
	<analytic>
		<title level="a" type="main">Tag recommendation for folksonomies oriented towards individual users</title>
		<author>
			<persName><forename type="first">M</forename><surname>Lipczak</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the ECML/PKDD 2008 Discovery Challenge Workshop, part of the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases</title>
				<meeting>the ECML/PKDD 2008 Discovery Challenge Workshop, part of the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b241">
	<analytic>
		<title level="a" type="main">Overview of the OKAPI projects</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">E</forename><surname>Robertson</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Documentation</title>
		<imprint>
			<biblScope unit="volume">53</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="3" to="7" />
			<date type="published" when="1997">1997</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b242">
	<analytic>
		<title level="a" type="main">Rsdc&apos;08: Tag recommendations using bookmark content</title>
		<author>
			<persName><forename type="first">M</forename><surname>Tatu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Srikanth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">D</forename><surname>Silva</surname></persName>
		</author>
		<ptr target="http://www.vanderwal.net/folksonomy.html" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the ECML/PKDD 2008 Discovery Challenge Workshop, part of the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases</title>
				<editor>
			<persName><forename type="first">T</forename><forename type="middle">V</forename><surname>Wal</surname></persName>
		</editor>
		<meeting>the ECML/PKDD 2008 Discovery Challenge Workshop, part of the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases</meeting>
		<imprint>
			<date type="published" when="2007">2008. February 2, 2007</date>
			<biblScope unit="page" from="96" to="107" />
		</imprint>
	</monogr>
	<note>Folksonomy coinage and definition</note>
</biblStruct>

<biblStruct xml:id="b243">
	<analytic>
		<title level="a" type="main">Information retrieval in folksonomies: Search and ranking</title>
		<author>
			<persName><forename type="first">A</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Jäschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Schmitz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the 3rd European Semantic Web Conference (ESWC)</title>
				<meeting>of the 3rd European Semantic Web Conference (ESWC)</meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b244">
	<analytic>
		<title level="a" type="main">Ht06, tagging paper, taxonomy, flickr</title>
		<author>
			<persName><forename type="first">C</forename><surname>Marlow</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Naaman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Boyd</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Davis</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">HYPERTEXT &apos;06</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
	<note>academic article, to read</note>
</biblStruct>

<biblStruct xml:id="b245">
	<analytic>
		<title level="a" type="main">Tag recommendations in folksonomies</title>
		<author>
			<persName><forename type="first">R</forename><surname>Jäschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Marinho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Schmidt-Thieme</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Lernen -Wissensentdeckung -Adaptivität (LWA) Workshop Proc</title>
				<imprint>
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b246">
	<analytic>
		<title level="a" type="main">Rsdc&apos;08: Tag recommendations using bookmark content</title>
		<author>
			<persName><forename type="first">M</forename><surname>Tatu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Srikanth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>D'silva</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the ECML PKDD Discovery Challenge (RSDC08)</title>
				<meeting>of the ECML PKDD Discovery Challenge (RSDC08)</meeting>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b247">
	<analytic>
		<title level="a" type="main">Tag recommendation for folksonomies oriented towards individual users</title>
		<author>
			<persName><forename type="first">M</forename><surname>Lipczak</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the ECML PKDD Discovery Challenge (RSDC08)</title>
				<meeting>of the ECML PKDD Discovery Challenge (RSDC08)</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b248">
	<monogr>
		<title level="m" type="main">Collaborative creation of communal hierarchical taxonomies in social tagging systems</title>
		<author>
			<persName><forename type="first">Paul</forename><surname>Heymann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Garcia-Molina</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2006">2006</date>
		</imprint>
		<respStmt>
			<orgName>Computer Science Department, Stanford University</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Technical report</note>
</biblStruct>

<biblStruct xml:id="b249">
	<analytic>
		<title level="a" type="main">Collaborative tagging and semiotic dynamics</title>
		<author>
			<persName><forename type="first">C</forename><surname>Cattuto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Loreto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Pietronero</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">PNAS</title>
		<imprint>
			<biblScope unit="volume">104</biblScope>
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b250">
	<analytic>
		<title level="a" type="main">The complex dynamics of collaborative tagging</title>
		<author>
			<persName><forename type="first">H</forename><surname>Halpin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Robu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Shepherd</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">WWW &apos;07: Proc. of the 16th int. conf. on World Wide Web</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b251">
	<analytic>
		<title level="a" type="main">The recurrence dynamics of social tagging</title>
		<author>
			<persName><forename type="first">D</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Mao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Li</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">WWW &apos;09: Proc. of the 18th int. conf. on World wide web</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b252">
	<analytic>
		<title level="a" type="main">Usage patterns of collaborative tagging systems</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">A</forename><surname>Golder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">A</forename><surname>Huberman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">J. of Information Science</title>
		<imprint>
			<biblScope unit="volume">32</biblScope>
			<biblScope unit="issue">2</biblScope>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b253">
	<analytic>
		<title level="a" type="main">Can all tags be used for search?</title>
		<author>
			<persName><forename type="first">K</forename><surname>Bischoff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">S</forename><surname>Firan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Nejdl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Paiu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CIKM &apos;08: Proc. of the 17th ACM conf. on Information and knowledge management</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
	<note>References</note>
</biblStruct>

<biblStruct xml:id="b254">
	<analytic>
		<title level="a" type="main">RSDC&apos;08: Tag Recommendations using Bookmark Content</title>
		<author>
			<persName><forename type="first">Marta</forename><surname>Tatu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Munirathnam</forename><surname>Srikanth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Thomas D'</forename><surname>Silva</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of ECML PKDD Discovery Challenge (RSDC 08)</title>
				<meeting>of ECML PKDD Discovery Challenge (RSDC 08)</meeting>
		<imprint>
			<biblScope unit="page" from="96" to="107" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b255">
	<analytic>
		<title level="a" type="main">Tag Recommendation for Folksonomies Oriented towards Individual Users</title>
		<author>
			<persName><forename type="first">Marek</forename><surname>Lipczak</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of ECML PKDD Discovery Challenge(RSDC 08)</title>
				<meeting>of ECML PKDD Discovery Challenge(RSDC 08)</meeting>
		<imprint>
			<biblScope unit="page" from="84" to="95" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b256">
	<analytic>
		<title level="a" type="main">BibSonomy: A Social Bookmark and Publication Sharing System</title>
		<author>
			<persName><forename type="first">Andreas</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Robert</forename><surname>Jaschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Chiristoph</forename><surname>Schmitz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gerd</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. the First Conceptual Structures Tool Interoperability Workshop at the 14 th Int. Conf. on Conceptual Structures</title>
				<meeting>the First Conceptual Structures Tool Interoperability Workshop at the 14 th Int. Conf. on Conceptual Structures<address><addrLine>Aalborg</addrLine></address></meeting>
		<imprint>
			<publisher>Aalborg Universitetsforlag</publisher>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="87" to="102" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b257">
	<analytic>
		<title level="a" type="main">FolkRank: A Ranking Algorithm for Folksonomies</title>
		<author>
			<persName><forename type="first">Andreas</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Robert</forename><surname>Jaschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Christoph</forename><surname>Schmitz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gerd</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of FGIR</title>
				<meeting>of FGIR</meeting>
		<imprint>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b258">
	<analytic>
		<title level="a" type="main">Tag Recommendations in Folksonomies</title>
		<author>
			<persName><forename type="first">Robert</forename><surname>Jaschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Leandro</forename><surname>Marinho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Andreas</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lars</forename><surname>Schmidt-Thieme</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gred</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Knowledge Discovery in Database: PKDD 2007, 11 th European Conference on Principles and Practice of Knowledge Discovery in Database</title>
				<meeting><address><addrLine>Warswa, Poland</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2007">September 17-21, 2007. 2007</date>
			<biblScope unit="volume">4702</biblScope>
			<biblScope unit="page" from="506" to="514" />
		</imprint>
	</monogr>
	<note>Proceedings</note>
</biblStruct>

<biblStruct xml:id="b259">
	<analytic>
		<title level="a" type="main">Probabilistic Author-Topic Models for Information Discovery</title>
		<author>
			<persName><forename type="first">Mark</forename><surname>Steyvers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Padhraic</forename><surname>Smyth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Thomas</forename><surname>Griffiths</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of KDD&apos;04</title>
				<meeting>of KDD&apos;04</meeting>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b260">
	<analytic>
		<title level="a" type="main">A Topic Modeling Approach and its Integration into the Random Walk Framework for Academic Search</title>
		<author>
			<persName><forename type="first">Jie</forename><surname>Tang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ruoming</forename><surname>Jin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jing</forename><surname>Zhang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceeding of 2008 IEEE international Conference on Data Mining(ICDM</title>
				<meeting>eeding of 2008 IEEE international Conference on Data Mining(ICDM</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="1055" to="1060" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b261">
	<analytic>
		<title level="a" type="main">and Larry Birnbaum TagAssist: Automatic tag suggestion for blog posts</title>
		<author>
			<persName><forename type="first">C</forename><surname>Sanjay</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Sara</forename><forename type="middle">H</forename><surname>Sood</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kristian</forename><forename type="middle">J</forename><surname>Owsley</surname></persName>
		</author>
		<author>
			<persName><surname>Hammond</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. the International Conference on Weblogs and Social Media(ICWSM</title>
				<meeting>the International Conference on Weblogs and Social Media(ICWSM</meeting>
		<imprint>
			<date type="published" when="2007">2007. 2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b262">
	<analytic>
		<title level="a" type="main">Autotag: a collaborative approach to automated tag assignment for weblog posts</title>
		<author>
			<persName><forename type="first">Gilad</forename><surname>Mishne</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">WWW&apos;06: Proceedings of the 15 th international conference on World Wide Web</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM Press</publisher>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="953" to="954" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b263">
	<analytic>
		<title level="a" type="main">RSDC&apos;08: Tag Recommendations using Bookmark Content</title>
		<author>
			<persName><forename type="first">Marta</forename><surname>Tatu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Munirathnam</forename><surname>Srikanth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Thomas D'</forename><surname>Silva</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of ECML PKDD Discovery Challenge (RSDC 08)</title>
				<meeting>of ECML PKDD Discovery Challenge (RSDC 08)</meeting>
		<imprint>
			<biblScope unit="page" from="96" to="107" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b264">
	<analytic>
		<title level="a" type="main">Tag Recommendation for Folksonomies Oriented towards Individual Users</title>
		<author>
			<persName><forename type="first">Marek</forename><surname>Lipczak</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of ECML PKDD Discovery Challenge(RSDC 08)</title>
				<meeting>of ECML PKDD Discovery Challenge(RSDC 08)</meeting>
		<imprint>
			<biblScope unit="page" from="84" to="95" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b265">
	<analytic>
		<title level="a" type="main">BibSonomy: A Social Bookmark and Publication Sharing System</title>
		<author>
			<persName><forename type="first">Andreas</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Robert</forename><surname>Jaschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Chiristoph</forename><surname>Schmitz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gerd</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. the First Conceptual Structures Tool Interoperability Workshop at the 14 th Int. Conf. on Conceptual Structures</title>
				<meeting>the First Conceptual Structures Tool Interoperability Workshop at the 14 th Int. Conf. on Conceptual Structures<address><addrLine>Aalborg</addrLine></address></meeting>
		<imprint>
			<publisher>Aalborg Universitetsforlag</publisher>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="87" to="102" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b266">
	<analytic>
		<title level="a" type="main">FolkRank: A Ranking Algorithm for Folksonomies</title>
		<author>
			<persName><forename type="first">Andreas</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Robert</forename><surname>Jaschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Christoph</forename><surname>Schmitz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gerd</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of FGIR</title>
				<meeting>of FGIR</meeting>
		<imprint>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b267">
	<analytic>
		<title level="a" type="main">Tag Recommendations in Folksonomies</title>
		<author>
			<persName><forename type="first">Robert</forename><surname>Jaschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Leandro</forename><surname>Marinho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Andreas</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lars</forename><surname>Schmidt-Thieme</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gred</forename><surname>Stumme</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Knowledge Discovery in Database: PKDD 2007, 11 th European Conference on Principles and Practice of Knowledge Discovery in Database</title>
				<meeting><address><addrLine>Warswa, Poland</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2007">September 17-21, 2007. 2007</date>
			<biblScope unit="volume">4702</biblScope>
			<biblScope unit="page" from="506" to="514" />
		</imprint>
	</monogr>
	<note>Proceedings</note>
</biblStruct>

<biblStruct xml:id="b268">
	<analytic>
		<title level="a" type="main">and Larry Birnbaum TagAssist: Automatic tag suggestion for blog posts</title>
		<author>
			<persName><forename type="first">C</forename><surname>Sanjay</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Sara</forename><forename type="middle">H</forename><surname>Sood</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kristian</forename><forename type="middle">J</forename><surname>Owsley</surname></persName>
		</author>
		<author>
			<persName><surname>Hammond</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. the International Conference on Weblogs and Social Media(ICWSM</title>
				<meeting>the International Conference on Weblogs and Social Media(ICWSM</meeting>
		<imprint>
			<date type="published" when="2007">2007. 2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b269">
	<analytic>
		<title level="a" type="main">Autotag: a collaborative approach to automated tag assignment for weblog posts</title>
		<author>
			<persName><forename type="first">Gilad</forename><surname>Mishne</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 15 th international conference on World Wide Web</title>
				<meeting>the 15 th international conference on World Wide Web<address><addrLine>New York, NY, USA; New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM Press</publisher>
			<date type="published" when="2006">2006. 2006</date>
			<biblScope unit="page" from="953" to="954" />
		</imprint>
	</monogr>
	<note>WWW&apos;06</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
