<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">The Evolution of the Semantic Web</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">John</forename><surname>Cardiff</surname></persName>
							<email>john.cardiff@ittdublin.ie</email>
						</author>
						<author>
							<persName><forename type="first">Rodrigo</forename><surname>Alarcón</surname></persName>
							<email>ralarconm@iingen.unam.mx</email>
						</author>
						<author>
							<persName><forename type="first">Octavio</forename><surname>Sánchez</surname></persName>
							<email>osanchezv@iingen.unam.mx</email>
						</author>
						<author>
							<persName><forename type="first">Víctor</forename><surname>Mijangos</surname></persName>
							<email>vmijangosc@iingen.unam.mx</email>
						</author>
						<author>
							<persName><forename type="first">Arturo</forename><surname>Velasco</surname></persName>
							<email>arturoezak@hotmail.com</email>
						</author>
						<author>
							<persName><forename type="first">María</forename><forename type="middle">J</forename><surname>Somodevilla</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Ivo</forename><forename type="middle">H</forename><surname>Pineda</surname></persName>
							<email>ipineda@cs.buap.mx</email>
						</author>
						<author>
							<persName><forename type="first">David</forename><surname>Pinto</surname></persName>
							<email>dpinto@cs.buap.mx</email>
						</author>
						<author>
							<persName><forename type="first">Mireya</forename><surname>Tovar</surname></persName>
							<email>mtovar@cs.buap.mx</email>
						</author>
						<author>
							<persName><forename type="first">Beatriz</forename><surname>Beltrán</surname></persName>
							<email>bbeltran@cs.buap.mx</email>
						</author>
						<author>
							<persName><forename type="first">Darnes</forename><surname>Vilariño</surname></persName>
							<email>darnes@cs.buap.mx</email>
						</author>
						<author>
							<persName><forename type="first">Héctor</forename><surname>Furlog</surname></persName>
						</author>
						<author>
							<persName><forename type="first">María</forename><surname>Elena</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Franco</forename><surname>Carcedo</surname></persName>
							<email>fcarcedo@hotmail.com</email>
						</author>
						<author>
							<affiliation key="aff0">
								<orgName type="laboratory">Social Media Research Group</orgName>
								<orgName type="institution">Institute of Technology Tallaght</orgName>
								<address>
									<settlement>Dublin</settlement>
									<country key="IE">Ireland</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff1">
								<orgName type="department">Grupo de Ingeniería Lingüística</orgName>
								<orgName type="institution">Universidad Nacional Autónoma de México Basamento de la Torre de Ingeniería</orgName>
								<address>
									<addrLine>Ciudad Universitaria</addrLine>
									<region>México, D.F</region>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff2">
								<orgName type="department">Facultad de Ciencias de la Computación</orgName>
								<orgName type="institution">Universidad Autónoma de Puebla</orgName>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff3">
								<orgName type="department">Faculty of Computer Science</orgName>
								<orgName type="institution">BUAP 14 Sur &amp; Av. San Claudio</orgName>
								<address>
									<addrLine>CU, Edif. 104C</addrLine>
									<postCode>72570</postCode>
									<settlement>Puebla</settlement>
									<region>Mexico</region>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff4">
								<orgName type="department">Facultad de Ciencias de la Computación</orgName>
								<orgName type="institution">Benemerita Universidad Autonoma de Puebla</orgName>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff5">
								<orgName type="department">BUAP -FCC</orgName>
								<address>
									<addrLine>Ciudad Universitaria, Apartado Postal J-32</addrLine>
									<settlement>Puebla</settlement>
									<region>Pue. México</region>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff6">
								<orgName type="institution">Université Claude Bernard Lyon</orgName>
								<address>
									<addrLine>1 Bâtiment Nautibus 43, Boulevard du 11 novembre</addrLine>
									<postCode>1918, 69622</postCode>
									<settlement>Villeurbanne Cedex</settlement>
									<country key="FR">France</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff7">
								<orgName type="institution">OMAR.NINO</orgName>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff8">
								<orgName type="institution">Universidad Politecnica de Cataluña</orgName>
								<address>
									<addrLine>Jordi Girona</addrLine>
									<postCode>1-3</postCode>
									<settlement>Barcelona</settlement>
									<country key="ES">España</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff9">
								<orgName type="institution" key="instit1">ICUAP</orgName>
								<orgName type="institution" key="instit2">RIFCCBUAP</orgName>
								<address>
									<addrLine>Ciudad Universitaria</addrLine>
									<postCode>72570</postCode>
									<settlement>Puebla</settlement>
									<region>CA</region>
									<country key="MX">Mexico</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">The Evolution of the Semantic Web</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">61F278DD43B89F59113E78D27799B337</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T21:46+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Semantic Web</term>
					<term>Ontology</term>
					<term>Web of Data Wikipedia mining</term>
					<term>Wikipedia infobox</term>
					<term>ontology construction</term>
					<term>semantic relation extraction</term>
					<term>information extraction</term>
					<term>natural language processing automatic generation of taxonomies</term>
					<term>dynamic taxonomies</term>
					<term>faceted classification</term>
					<term>faceted taxonomies</term>
					<term>information recovery</term>
					<term>navigation trees Semantic Web</term>
					<term>verbal language</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The Semantic Web offers an exciting promise of a world in which computers and humans can cooperate effectively with a common understanding of the meaning of data. However, in the decade since the term has come into widespread usage, Semantic Web applications have been slow to emerge from the research laboratories. In this paper, we present a brief overview of the Semantic Web vision and the underlying technologies. We describe the advances made in recent years and explain why we believe that Semantic Web technology will be the driving force behind the next generation of Web applications.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>INTRODUCTION</head><p>The World Wide Web (WWW) was invented by Tim Berners Lee in 1989, while he was working at the European Laboratory for Particle Physics (CERN) in Switzerland. It was conceived as a means to allow physicists working in different countries to communicate and to share documentation more efficiently. He wrote the first browser and Web server, allowing hypertext documents to be stored, retrieved and viewed.</p><p>The Web added two important services to the internet -it provided a very convenient means for us to retrieve and view information -we can then see the web as a vast document store in which we retrieve documents (web pages) by typing in their address into a web browser. Secondly, it provided a language called HTML, which describes to computers how to display documents written in this language. Documents, or web pages, are accessed by a unique identifier called a Uniform Resource Locator (URL) and are accessed using a Web browser. Within a short space of time, the WWW had become a popular infrastructure for sharing information, and as the volume of information increased its use became increasingly widespread.</p><p>Although the web provides the infrastructure for us to publish and retrieve documents, the HTML language defines only the visual characteristics, ie. how the documents are to be presented on a computer screen to the user. It is up to the user who requested the document to interpret the information it contains. This seems counterintuitive, as we normally think of computers as the tools to perform the more complex tasks, making life easier for humans. The problem is that within HTML there is no consideration of the meaning of the document, they are not represented in a way that allows interpretation of their information content by computers.</p><p>If computers could interpret the content of a web page, a lot of exciting possibilities would arise. Information could be exchanged between machines, automated processing and integration of data on different sites could occur. Fundamentally, they could improve the ways in which they can retrieve and utilise the information for us because they would have an understanding of what we are interested in. This is where the Semantic Web fits into the picturetoday's web (the "syntactic" web) is about documents whereas the semantic web is about "things" -concepts we are interested in (people, places, events etc.), and the relationships between these concepts.</p><p>The Semantic Web vision envisages advanced management of the information on the internet, allowing us to pose queries rather than browse documents, to infer new knowledge from existing facts, and to identify inconsistencies. Some of the advantages of achieving this goal include <ref type="bibr">[4]</ref>:</p><p> The ability to locate information based on its meaning, eg. knowing when two statements are equivalent, or knowing that a reference to a person in different web pages are referring to the same individual.  Integrating information across different sources − by creating mappings across application and terminological boundaries we can identify identical or related concepts,  Improving the way in which information is presented to a user, eg. aggregating information from different sources, removing duplicates, and summarising the data.</p><p>While the technologies to enable the development of the Semantic Web were in place from the conception of the web, a seminal article by Tim Berners-Lee, James Hendler and Ora Lassila <ref type="bibr">[1]</ref> published in Scientific American in 2001 provided the impetus for research and development to commence. The authors described a world in which independent applications could cooperate and share their data in a seamless way to allow the user to achieve a task with minimal intervention. Central to this vision is the ability to "unlock" data that is controlled by different applications and make it available for use by other applications. Much of this data is already available on the Web, for example we can access our bank statements, our diaries and our photos online. But the data is controlled by proprietary applications. The Semantic Web vision is to publish this data in a sharable form − we could integrate the items of our bank statements into our calendar so that we could see what transactions we made on that day, or include photos so that we could see what we were doing at that time.</p><p>However, eight years after publication of this article, we are still some distance realising this vision. In this paper, present an overview of the Semantic Web. We explain why progress has been slow and the reasons we believe this to be about to change.</p><p>The paper is organized as follows. In Section II we describe the problems we face when trying to extract meaning from the web as it is today. Section III presents a brief overview of the technologies underlying the Semantic Web. In Section IV we give an overview of the gamut of typical Semantic Web applications and Section V introduces the Linking Open Data project. Finally, we present our conclusions and look to the future in Section VI.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>II. THE PROBLEM WITH THE "SYNTACTIC WEB"</head><p>In Figure <ref type="figure" target="#fig_0">1</ref> we see a "typical" web page written in HTML which we will use to exemplify some of the drawbacks of the traditional web. This page lists the keynote speeches which took place at the 2009 World Wide Web conference 1 .</p><p>To the reader, the content of the page can be interpreted intuitively. We can read the titles of the speeches, the names of the speakers and the time and dates at which they take place. Furthermore, by familiarity with browser interaction paradigms, we can realize that by following a hyperlink we can retrieve information about concepts related to the conference (authors, sponsors, attendees etc.). In this example, by following the hyperlink labelled "Sir Tim Berners-Lee" we will retrieve a document containing information about the person of this name. We intuitively assign a meaning -perhaps "has-homepage" -to the hyperlink, allowing us to assimilate the information presented to us.</p><p>A web browser cannot assign any to these links we see in this page − a hyperlink is simply a link from one document to another and the interpretation of the meaning of the link (and of the documents themselves!) is a task for the human reader. All that can be inferred automatically is that some undefined association between the two documents exists.</p><p>1 http://www2009.org/keynote.html The problems are even more clear when we consider the nature of keyword-based browsing. While search engines such as Google and Yahoo! are clearly very good at what they do, we frequently are presented with a vast number of results, many (most?) of which will be irrelevant to our search. Semantically similar items will not be retrieved (for instance a search for "movie" will not retrieve results where the word "film" was used). And most significantly, the result set is a collection of individual web pages. Our tasks often require access to multiple sites (such as when we book a holiday), and so it is our responsibility to formulate a sequence of queries to retrieve the individual web pages, each one of which performs part of the task at hand.</p><p>There are two potential ways to deal with this problem. One approach is to take the web as it is currently implemented, and to use Artificial Intelligence techniques to analyze the content of web pages in order to provide an interpretation of its meaning. This approach however would be prone to error and would require validation. Furthermore, the rate at which the web is growing would render it practically impossible to achieve.</p><p>The other approach is to represent the web pages in a form in which we can represent and interpret the data they contain. If there is a common representation to express the meaning of the data on the web, we can then develop languages, reasoners, and applications which can exploit this representation. This is the approach of the Semantic Web.</p><p>III. SEMANTIC WEB TECHNOLOGIES The Semantic Web describes a web of data rather than documents. And just as we need common formats and standards to be able to retrieve documents from computers all over the world, we need common formats for the representation and integration of data. We also need languages that allow us to describe how this data relates to real world objects and to reason about the data. The famous "Layer Cake" <ref type="bibr">[10]</ref> diagram, shown in Figure <ref type="figure" target="#fig_1">2</ref>, gives an overview of the hierarchy of the principal languages and technologies, each one exploiting the features of the levels beneath it. It also reinforces the fact that the Semantic Web is not separate from the existing web, but is in fact an extension of its capabilities.</p><p>In this section, we summarize and discuss the key aspects shown in the Layer Cake diagram. Firstly we describe the core technologies: the languages RDF and RDFS. Next we describe the higher level concepts, focusing in particular on the concept of the ontology which is at the heart of the Semantic Web infrastructure. Finally we examine the trends and directions of the technology. For further information on the concepts presented in this section, the reader is referred to a more detailed work (eg. <ref type="bibr">[4]</ref>, <ref type="bibr">[5]</ref>). </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A. The Core Technologies: RDF and RDFS</head><p>What HTML is to documents, RDF (Resource Description Framework) is to data. It is a W3C standard 2 based on XML which allows us to make statements about objects. It is a data model rather than a language -we can say that an object possesses a particular property, or that it has a named relationship with another object. RDF statements are written as triples: a subject, predicate and object.</p><p>By way of example, the statement 2 www.w3.org/RDF/ "The Adventures of Tom Sawyer" was written by Mark Twain could be expressed in RDF by a statement such as &lt;rdf:Description rdf:about=www.famouswriters.org/twain/mark&gt; &lt;s:hasName&gt;Mark Twain&lt;/s:hasName&gt; &lt;s:hasWritten rdf:resource= www.books.org/ISBN0001047&gt; &lt;/rdf:Description&gt; At first glance it may appear that this information could be equally well represented using XML. However XML makes no commitment on which words should be used to describe a given set of concepts. In the above example we have a property entitled "hasWritten", but this could equally have been "IsAuthorOf" or another such variant. So, XML is suitable for closed and stable domains, rather than for sharable web resources.</p><p>The statements we make in RDF are unambiguous and have a uniform structure. Concepts are each identified by a Universal Resource Identifer (URI) which allows us to make statements about the same concept in different applications. This provides the basis for semantic interoperability, allowing us to distinguish between ambiguous terms (for instance an address could be a geographical location, or a speech) and to define a place on the web at which we can find the definition of the concept.</p><p>To describe and make general statements collectively about groups of objects (or classes), and to assign properties to members of these groups we use RDF Schema, or RDFS <ref type="foot" target="#foot_0">3</ref> . RDFS provides a basic object model, and enables us to describe resources in terms of classes, properties, and values. Whereas in RDF we spoke about specific objects such as '"The Adventures of Tom Sawyer" and "Mark Twain", in RDFS we can make general statements such as "A book was written by an author" This could be expressed in RDFS as &lt;rdf:Property rdf:ID="HasWritten" &lt;rdfs:domain rdf:resource="#author"\&gt; &lt;rdfs:range rdf:resource="#book"\&gt; &lt;\rdf:Property&gt; An expansion of these examples, and the relationship between the graphical representations of RDF and RDFS is shown in Figure <ref type="figure" target="#fig_2">3</ref>. B. Ontologies and Reasoning RDF and RDFS allow us to describe aspects of a domain, but the modelling primitives are too restrictive to be of general use. We need to be able to describe the taxonomic structure of the domain, to be able to model restrictions or constraints of the domain, and to be able to state and reason over a set of inference rules associated with the domain. We need to be able to describe an ontology of our domain.</p><p>The term ontology originated in the sphere of philosophy, where it signified the nature and the organisation of reality, ie. concerning the kinds of things that exist, and how to describe them. Our definition within Computer Science is more specific, and the most commonly cited definition has been provided to us by Tom Gruber in <ref type="bibr">[6]</ref>, where he defines an ontology as "an explicit and formal specification of a conceptualization". In other words, an ontology provides us with a shared understanding of a domain of interest. The fact that the specification is formal means that computers can perform reasoning about it. This in turn will improve the accuracy of searches, since a search engine can retrieve data regarding a precise concept, rather than a large collection of web pages based on keyword matching.</p><p>In relation to the Semantic Web, for us to share, reuse and reason about data we must provide a precise definition of the ontology, and represent it in a form that makes it amenable to machine processing. An ontology language should ideally extend existing standards such as XML and RDF/S, be of "adequate" expressive power, and provide efficient automated reasoning support. The most widely used ontology language is the "Web Ontology Language", which curiously has the acronym "OWL" 4 . Along with RDF/S, OWL is a W3C standard and augments RDFS with additional constraints such as localised domain and range constraints, cardinality and existence constraints, and transitive, inverse, and symmetric properties.</p><p>Adding a reasoning capability to an ontology language is tricky since there will be a trade-off between efficiency and expressiveness. Ultimately it depends on the nature and requirements of the end application, and it is for this reason that OWL offers three sublanguages, 4 www.w3.org/2004/OWL  OWL Lite supports only a limited subset of OWL constructs and is computationally efficient,  OWL DL is based on a first order logic called Description Logic,  OWL Full offers the full compatibility with RDFS but at the price of computational tractability.</p><p>Examples of applications which could require very different levels of reasoning capabilities are described in the following section.</p><p>The top layers of the layer cake have received surprising little attention considering that they are crucial to successful deployment of Semantic Web applications. The proof layer involves the actual deductive process, representation of proofs, and proof validation. It allows applications to inquire why a particular conclusion has been reached, ie. they can give proof of their conclusions. The trust layer provides authentication of identity and evidence of the trustworthiness of data and services. It is supported through the use of digital signatures, recommendations by trusted agents, ratings by certification agencies etc.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>C. Recent Trends and Technological Developments</head><p>As with any maturing technology, the architecture will not remain static. In 2006 Tim Berners Lee suggested an update to the layer cake diagram <ref type="bibr">[2]</ref> which is shown in Figure <ref type="figure" target="#fig_3">4</ref>, however this is just one of several proposed refinements. Some of the new features and languages which include the following.</p><p>Rules and Inferencing Systems. Alternative approaches to rule specification and inferencing are being developed. RIF (Rules Interchange Format) is a language for representing rules on the Web and for linking different rule-based systems together. The various formalisms are being extended in order to capture causal, probabilistic and temporal knowledge.</p><p>Database Support for RDF. As the volume of RDF data increases, it is necessary to provide the means to store, query and reason efficiently over the data. Database support for RDF and OWL is now available from Oracle (although at present the focus is on storage, rather than inferencing capabilities). Other open source products include 3Store <ref type="foot" target="#foot_1">5</ref>and Jena <ref type="foot" target="#foot_2">6</ref> . The specification of a query language for RDF, SPARQL, was adopted by the W3C in 2008.</p><p>RDF Extraction. The language GRDDL: ("Gleaning Resource Descriptions from Dialects of Languages") identifies when an XML document contains data compatible with RDF and provides transformations which can extract the data. Considering the volume of XML data available on the web, a means of converting this to RDF is clearly highly desirable.</p><p>Ontology Language Developments. The OWL language was adapted as a standard in 2004. In 2007, work began on the definition of a new version, OWL 2 which includes easier query capabilities and efficient reasoning algorithms scaled to large datasets. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>IV. THE SPECTRUM OF APPLICATIONS</head><p>Even though Semantic Web technology is in its infancy, there are a wide range of applications in existence. In this section we give a brief overview of some typical application areas.</p><p>E-Science Applications. Typically e-science describes scenarios involving large data collections requiring computationally intensive processing, and where the participants are distributed across the world. An infrastructure whereby scientists from different disciplines are able to share their insights and results is seen as critical, particularly when we consider the availability of large volumes of data becoming available online. The Gene Ontology 7 is a project aimed at standardizing the representation of genes across databases and species. Perhaps the most famous e-science project is the Human Genome Project 8 which identified the genes in human DNA and which includes over 500 datasets and tools. The International Virtual Observatory Alliance 9 makes available astronomical data from a number of digital archives.</p><p>Interoperation of Digital Libraries. Institutions such as libraries, universities, and museums have vast inventories of materials which are increasingly becoming available online. These systems are implemented using a range of different technologies, and although their aims are similar it is a huge challenge to enable the different institutions to access each other's catalogues. Ontologies are useful for providing shared descriptions of the objects, and ontology mapping techniques are being applied to achieve semantic interoperability <ref type="bibr">[3]</ref>.</p><p>Travel Information Systems. The goal of building an application which would allow a user to seamlessly book and plan the various elements of a trip (flights, hotel, car hire etc.) is highly desirable. Ontologies again could be used to arrive at a common understanding of terminology. The Open Travel Alliance is building XML based specifications which allow for the interchange of messages between companies. While this is a first step, an agreed ontology would be needed in order to achieve any meaningful interoperation.</p><p>Although many potential applications can be identified, there are less deployed at this time than we might expect. One possible reason is the lack of a common understanding of what the Semantic Web can offer, and more particularly what the role of ontology. At one end of the spectrum we find applications which take the "traditional", or AI view of inferencing, in which accuracy is paramount. Such applications arise in combinatorial chemistry for example, in which vast quantities of information on chemicals and their properties are analysed in order to identify useful new drugs. By coding the required drug's properties as assertions will reduce the number of samples which need to be constructed and manually analyzed by orders of magnitude.</p><p>In cases such as these, the time taken to perform the inferencing is less important, since the trade-off will be a large reduction in the samples to be analyzed.</p><p>At the other end of the spectrum, we have "data centric" web applications which require a swift response to the user.</p><p>Examples of this type of application include social network recommender systems such as Twine<ref type="foot" target="#foot_3">10</ref> which make use of ontologies to recommend their users to other individuals who may be of interest to them. While it is clear that a response must be generated for the user within a few seconds, we can observe too that there can be no logical proof of correctness and soundness of the answers generated in this type of case! Accordingly, the level of inferencing required in this type of application is minimal.</p><p>V. THE FUTURE: A WEB OF DATA?</p><p>While we have stated that the Semantic Web focuses on data in contrast to the document centric view of the traditional web, this is not the complete picture. In order to realize value from putting data on the web, links need to be made in order to create a "web of data". Instead of having a web with pages that link to each other, we can have (with the same infrastructure) a data model with information on each entity distributed over the web.</p><p>The Linking Open Data <ref type="bibr">[3]</ref> project aims to extend the collections of data being published on the web in RDF format and to create links between them. In a sense, this is analogous to traditional navigation between hypertext documents where the links are now the URIs contained in the RDF statements. Search engines could then query, rather than browse this information.</p><p>In a recent talk at the TDC 2009 conference 11 , Tim Berners Lee gave a powerful motivation example for the project: scientists investigating the drug discovery for Alzheimer's disease needed to know which proteins were involved in signal transduction and were related to pyramidal neurons. Searching on Google returned 223,000 hits, but no document provided the answer as nobody had asked the question before. Posing the same question to the linked data produces 32 hits, each of which is a protein meeting the specified properties. The knowledge base is constructed by analyzing the different types of structured information, such as the "infoboxes", tables, pictures etc.  The DBLP Bibliography, which contains bibliographic information of academic papers,  Geonames, which contains RDF descriptions of 6.5 million geographical features. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>VI. LOOKING AHEAD AND CONCLUSIONS</head><p>So where is the Semantic Web? In a 2006 article <ref type="bibr">[11]</ref>, Tim Berners Lee agreed that the vision he described in the The last five years have seen Semantic Web applications move from the research labs to the marketplace. While the use of ontologies has been flourishing in niche areas such as e-science for a number of years a recent survey by Hendler <ref type="bibr">[7]</ref> shows a marked increase in the number of commercially focused semantic web products. The main industrial players are starting to take the technology more seriously. In August 2008, Microsoft bought Powerset, a semantic search engine, for a reported $100m.</p><p>As we have discussed, the "chicken and egg" dilemma is resolving itself with tens of billions of RDF triples now available on the web, and this number is continuing to increase exponentially.</p><p>Also, it is becoming easier for companies to enter the market of Semantic Web applications. There are now a wide range of open source applications such as Protégé 13 and Kowari 14 which provide building blocks for application development, making it more cost effective to develop Semantic Web products.</p><p>Some observers argue that the Semantic Web has failed to deliver its promise, arguing instead that the Web 2.0 genre of applications signifies the way forward. The Web 2.0 approach has made an enormous impact in recent years, but these applications could be developed and deployed more rapidly as their designers did not have the inconvenience of standards to adhere to. In this article we have demonstrated the steady infiltration from the research lab to the marketplace being made by the Semantic Web over the last decade. As the standards mature and the web of data expands, we are confident that the Semantic Web vision is set to become a reality. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Wikipedia is a free encyclopedia of open content that has become an important resource towards the construction of the Semantic Web. Since it beginnings, in the year 2001, the English version has achieve more than 2 million of articles, while the Spanish version has around 480 thousand of articles. All of the content has been written and edited by volunteers from different countries in many different languages, and it is covered by GFDL (GNU Free Document License), which makes possible to freely use them.</p><p>One important thing about the structure of Wikipedia is the social control executed by the community, which is able to avoid the spam, the nonsense and other kind of vandalism that is recurrent on some media sites. Besides, this same control makes possible to constantly increase the quality and precision of the articles.</p><p>Inside Wikipedia, there is an entry called Wikipedia: Wikipedia in academic studies <ref type="foot" target="#foot_4">1</ref> , where it is possible to see the growth of academic interest in this encyclopedia. This interest is related to the use of Wikipedia on different academic studies and as a knowledge base for developing specific tools. On one hand, to mention a few, some works have focused on the social theme that represents Wikipedia <ref type="bibr">[1]</ref> [2], some other have denounced inherent problems presented on this kind of media sites <ref type="bibr">[3]</ref>, and others have obtained specific information and statistic data about the users <ref type="bibr">[4]</ref>. On the other hand, Wikipedia has become a useful resource for the extraction of definitions, name entity recognition, machine translation or semantic relation extraction <ref type="bibr">[5]</ref>. In this last field, Wikipedia represents a huge knowledge base that has made possible the developing of specific ontologies for the construction of the Semantic Web.</p><p>In this paper we present a work in process for the elaboration of an ontology of movies from Wikipedia on Spanish language. First we will briefly present an overview of some studies related to the use of Wikipedia for semantic relation extraction and ontology construction <ref type="bibr">(2)</ref>. Then we will explain the first step towards the elaboration of an ontology of movies (3). This step includes: a) the description of the so-called infobox, which is part of each movie of Wikipedia and contains specific data about the film (3.1); b) the specific relations to automatically extract (3.2); c) and our proposed XML schema to represent these relations <ref type="bibr">(3.3)</ref>. Finally, we will discuss our preliminary results and present the future work (4).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Wikipedia as a Semantic Knowledge Base</head><p>There is a growing interest of efforts to mine the information in Wikipedia for different purposes. As we have mentioned before, one of this interest is the extraction of semantic information that could be helpful on the process of giving more meaning to the Web. In Wikipedia, the meaning could be seen as the knowledge about things represented in different ways: definitions, descriptions, images, numeric data, etc. Furthermore, the meaning of each concept explained on the encyclopedia is related to the meaning of other concepts, which becomes a helpful semantic network to understand concepts on the field where they belong.</p><p>In this sense, Wikipedia represents a valuable source of knowledge to extract semantic information between concepts. A general overview of how Wikipedia could be used to extract concepts, relations, facts and descriptions can be found in <ref type="bibr">[6]</ref>. Here, the authors explain the use of Wikipedia for natural language processing, information extraction and ontology building.</p><p>In <ref type="bibr">[7]</ref>, the authors describe a methodology that uses the links between categories to mine specific relations. They analyze some measures to infer relations and try to provide a semantic scheme in order to improve the search capabilities and to give the users meaningful suggestions to edit articles. In the same context, in <ref type="bibr">[8]</ref> the authors use Wikipedia to develop a methodology for the automatic annotation of different semantic relations. This work is based on discovering lexical patterns that can be used to recognized specific relations between concepts. They evaluate the methodology by using a corpus and searching on it the relations founded in Wikipedia. Their results show that this kind of methodology could be a good starting point for automatic ontology construction.</p><p>The research presented in <ref type="bibr">[9]</ref> shows how hyperlinked pages are used to generate a domain hierarchy by means of ranking articles that are strongly linked. These articles become a domain corpus for the automatic construction of an ontology. The same goal of obtaining ontologies through Wikipedia is described in <ref type="bibr">[10]</ref>, where the authors apply machine learning techniques to improve the performance of a system that mines the infoboxes. Finally, in <ref type="bibr">[11]</ref> we can found another example of the use of Wikipedia for ontology construction, specifically for document classification. This is not, and does not pretend to be an extensive list of all the works made about semantic relation extraction or ontology construction from Wikipedia. Our main purpose is to state both the interest that has woken up in the area of extraction and organization of semantic information, and some of the automatic analyzes and procedures that are possible to develop taking into account Wikipedia's structure. Nevertheless, as we will see in this paper, this structure is often not well organized and makes it difficult to implement automatic processes.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Towards an Ontology of Movies</head><p>In order to develop an ontology of movies we have stated three main steps that can lead us to our purpose. The first one is to collect our input corpus from Wikipedia movies articles and the analysis of the infobox structure on them. After that, the second step is the delimitation and automatic extraction of specific semantic information. Finally, as a third step we consider the implementation of the extracted information into a XML schema that will conform the basis for another later annotation schema.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Movies infobox structure</head><p>The first step of our methodology was to conform a corpus from the articles of the films by year category. We use the categories tree option to find a list of the movies titles from the year 1892 to 2008 2 . After that, we use the export pages option to retrieve all the articles of this list. We found a total of 5,561 articles, where the opening and closing infoboxes tags ({{Fields…}}) was on 5,092 cases. This late number represents the total of articles from our corpus.</p><p>After that, we analyze the infobox of each entry. The infobox is a resource used on Wikipedia to summarize and group the information about specific data on some articles. In general words, its purpose is to make the information on a more available format and it can be use as a resource to other applications.</p><p>In Spanish language, there are 49 proposed fields for the infobox, where only two are consider as required: film title and original title. The infobox will be framed in {{Fields…}}, and each field inside will be preceded by a vertical bar "|" and followed by an equal sign "=" and the specific information. Fields without descriptions will remain empty after the equal sign. That means it will have the following structure: The whole fields used in the movies infoboxes from Wikipedia in Spanish can be found in table 1.</p><formula xml:id="formula_0">| Field =</formula><p>Table1.InfoboxtemplateinSpanish. From the table above we can see the different kind of information that the fields can introduce. We see information about dirección (direction), estreno (premiere), idioma (language, language2, language3, etc.), as well as país (country country2, country3, etc.), IMDb (Internet Movie Data Base) or Filmaffinity links (external Web sites with movies information).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Fields</head><p>The 49 fields from this table are the suggested ones in the official Wikipedia movies infobox template. Nevertheless, in our corpus we found several empty fields. We automatically found a total of 94,584 fields occurrences, while 30,742 cases where empty (32.48% of the whole occurrences).</p><p>Furthermore, one of the problems presented in the infoboxes is the lack of standardization. Some of the elements established by Wikipedia are written in an indistinctive way by the authors of the articles, while others have typographical errors. For example, the field dirección (direction) appears also as director (director); the field título original (original title) can be found as título en España (title in Spain), título principal (main title), título traducido (translated title), among others. More complicated is the case of estreno (premiere), which presents variations like año (year), fecha (date), fecha de estreno (premiere date), or primera emisión (first emission).</p><p>Typos are another common lack of standardization. For the field género (genre) we can find mistakes like *gènero, *genero or *genro.</p><p>In the corpus we can also find the case of another fields that are not proposed in the original schema, such as asistente de artes marciales (martial arts assistant), calificación (qualification), premios (awards), Myspace, and so on. In this case we found a total of 205 non-official fields.</p><p>If we compare the schema in Spanish to the English one, we can notice that the latter infobox contains fewer fields, which probably allows to be more standardized at the moment to put it into practice. The fields of the movies infobox in English can be seen in table 2. Here we can observe a total of 22 fields, comparing to the 49 in the Spanish template. It is important to notice the fact that most of another languages follow a similar structure like the one described for English. There is a similar template to the English movies infobox in French Wikipedia, with some added elements like format, awards, and IMDb. In Italian, the infobox determines general fields for different genres of films: generic, animation or film a episodi (films conformed from several short films), with specific fields for each genre; while in German, the fields specifies a more generic data, i.e., title, original title, producer or cameraman.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table2.InfoboxtemplateinEnglish</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Fields</head><p>In infoboxes of different languages, the most common fields are title, director and premiere. There are also coincidences in other fields, for example music and photography. Between English and Spanish there is a coincidence in preceded_by and followed_by. Furthermore, in Spanish, as well as in French, there is the field of IMDb, while Italian or English do not include. However, in English links to IMDb or Allmovie can appear within the article as external links and not inside the template of the infobox. These external links are also a valuable information to extend the semantic data for an ontology, as they can add more information about the films that does not appears in Wikipedia, or be used to complete the empty fields of the infoboxes. Nevertheless, there is also no consistence between the occurrences of the tags with external links. In our corpus, the IMDb tag occurs approximately in the 80% of the articles, while Filmaffinity occurs around in the 5%.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Extracting specific relations data</head><p>Theoretically, the structure of the infoboxes contains information that should be exploited with relative easiness. We decide to automatically extract the title, original title, director, premiere year and genre, in order to generate a database with all of this information. Although, not all of this information is present on all the movies articles founded in the films by year category.</p><p>As we have mentioned before, there are some inconsistencies within the name of the fields, their completeness, or the way the authors write them. In the case of director field, we found it with complete information in the 5,092 occurrences of the articles with infoboxes, however the field genre occurs only in 4,499 of these cases. Taking into account that the inconsistencies of the metadata make more difficult the process of automatic relation extraction from the films information, we achieve to obtain the data through the process we hereby describe.</p><p>From our corpus, we find out that 5,092 articles contained at least one director, although the field name from many of them was not the same and a review had to be made in order to compile a list of ad hoc synonyms for searching this specific field. The synset was formed by dirección (direction), director (director) and dirigida (directed). Also, after the equal sign that should follow the name of the field, the kind of following blanks was not always the same. Sometimes there were tabs; some other, more than one simple space; and even other, without spaces. Many of the director's names are also entries of Wikipedia, so many users decided to establish links to their names, using the symbol "[[" followed by the name of the director and closing with "]]". This has the purpose of specifying to the wikiengine that there is a link: [[link to the article]]. But not all of them had those brackets, and it caused troubles while parsing the data with the aim of recovering the director's name of the film associated with the title of the entry.</p><p>The same problems were founded when we tried to mine the original title of the movie. Despite the fact that this field does appear in all the infoboxes, not all of them appear with information, which means that there are articles with the original title field empty. It does not contain information in 195 articles occurrences in the corpus.</p><p>With the premiere field it was also problematic to extract the information, because most of the films had different words to express the premiere year, for example año (year), fecha de estreno (premiere date) or *añoacceso (acces year). In this case we decide to mine only the año (year) and estreno (premiere) variants, because of the wide range of structural possibilities. We found that 23 films infoboxes do not contain a premiere year, sometimes it was in the title and sometimes were completely absent.</p><p>Other field we exploited was the one of género (genre), which also present some inconsistencies that could be attached to human errors at the time of transcribing the template. This field was empty on 593 occurrences in our corpus and is the more unused one.</p><p>Summarizing, we can find the number of occurrences for each field in table <ref type="table" target="#tab_14">3</ref>: From the table above we can see the three fields with empty information: premiere or year, original title and genre. The first one was empty only in 23 articles, while the last one in more tan 500 cases. It is important to mention that the title of the movies was not obtained from the infobox but directly from the XML given by the Wikipedia, mainly because it is well demarcated by the labels &lt;title&gt; &lt;/title&gt;; in the same way, we obtained the id used by the Wikipedia to identify each article.</p><formula xml:id="formula_1">Table3.</formula><p>Despite the inconsistencies and typos that make difficult the automatic process, in 4,499 cases all the information that we were trying to mine was complete. We consider that this number represents a good starting point to conform the basis of a first schema that could be later extended.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Proposed XML schema</head><p>With the data from the infoboxes that were exploited, we decided to generate a first XML scheme, which should give basic information about the film. This scheme can be expanded as we extend our extraction processes of the information contained in the Wikipedia articles.</p><p>To make this scheme, we decided to take director field as the root XML tag. The first tag will consist of the director's name. Taking into account that directors can have more than one film, we decided to introduce a filmography tag to include them. This last tag will include each film with title, original title, year and genre tags. On the opening film tag we added an attribute with Wikipedia's title id number. An example of the schema can be seen below. As we can see in this example, the root tag is &lt;director&gt;&lt;/director&gt;. It is followed by the director's name tag &lt;name&gt;&lt;/ name &gt;. At the same level there is the tag &lt;filmography&gt;&lt;/ filmography&gt;. This tag nests the film tag &lt;film wiki_id=""&gt;&lt;/film&gt;, which contains the relevant information of each film: &lt;title&gt;&lt;/title&gt;, &lt;original_title&gt;&lt;/original_title&gt;, &lt;year&gt;&lt;/year&gt; and &lt;genre&gt; &lt;/genre&gt;.</p><p>Based on the XML scheme, relational databases can be generated to manipulate the information that we have considered at this first stage of the ontology construction. As we have said, this is not the full final scheme because as more data is extracted, the more can be added. This scheme is currently based on Wikipedia films articles in Spanish language, however it can be extended to fit another kind of relevant information, for example the country, external links (IMDb) or the id of directors or genre from Wikipedia. Furthermore, it will be possible to use this scheme in order to exploit Wikipedia in other languages, which could make possible to fill the empty fields in one language by relating them with the information on another languages, as well as to make multilingual queries.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Conclusions and future work</head><p>Nowadays, Wikipedia can be explored with the aim of obtaining information on different ways. The information added in a manual way by the users is generally well organized and semi-structured. Also, many entries from Wikipedia have infoboxes with summarized specific information about the theme treated in the article. We have mentioned that the structure of Wikipedia has made possible to exploit the information in order to extract semantic data. The extraction of semantic relations is one of the growing interests aiming to the construction of the Semantic Web.</p><p>Even so the structure of Wikipedia, we have noticed some specific problems on automatically exploiting it. To summarize a few, there are: a) the fact that the field's names are not respected; b) typos by human errors; c) lack of information; and d) differences on the infobox structure between languages. The latter should not be seen as a problem, however it would be advantageous to have standard fields on different languages.</p><p>Aiming to the standardization idea, it would be useful that the Wikipedia's process of writing or editing an article use a check-bot to confirm the information of the infoboxes templates. Thus, the fields not belonging to the template would be alerted, as well as typos on the field names. Furthermore, the same check-bot could be used to seek the existing fields looking for inconsistencies in the infoboxes or the whole articles.</p><p>The work that we have presented here is a first approach towards the elaboration of an ontology of movies from the Wikipedia in Spanish. We have showed the kind of semantic relations that are possible to mine, as well as a first scheme to represent them. We are conscious that this scheme may well be improved for achieving a complete ontology of movies. The future work will include: a) to define a scheme to represent subject, relation and predicates between the extracted information, for example a RDF scheme; b) to implement this new scheme for making the information available and sharing it with systems dedicated to the construction of the Semantic Web; c) to develop a movie-ontology query system capable of retrieve the information on specific ways related to directors, titles, genres and years fields.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>The UFs or fixed expressions are expressions consisting of two or more words whose meaning can not be inferred from the union of the significance of each of the lexical elements that constitute it. Zuluaga A., describes two basic characteristics that have fixed expressions: idiomaticity, characteristics that are peculiar and unique to a specific language or sub-language, including some socio-cultural traits; and fixation, the property that has the expressions of being reproduced in the speech like previously defined combinations, i.e. they present certain order in their syntactic structure <ref type="bibr">[1]</ref>.</p><p>Other important characteristics that have the UFs are: high frequency of report of their constituent elements, absence of grammatical rules in the expressions and translation problems.</p><p>On the other hand, due to the lack of agreement between linguists to establish limits of research of the phraseology and terminology used in this area, we decided to follow Alberto Zuluaga's work <ref type="bibr">[1]</ref>.</p><p>Zuluaga carries out a classification of the UFs based on the actions of the expressions in the speech. In the first group, Zuluaga establishes the locutions like a stable combination of two or more terms that work as an element in sentences to level of lexeme or syntagm. Inside this classification (locutions) he separates those that are in use as grammatical instruments and the expressions that possess semantic sense (lexical units). The subset of the UFs object of study in this work are the locutions and verbal syntagms whom belong to the units with lexical sense. The verbal locution is equivalent to lexemes, e.g.: pasar a mejor vida (to die) or echar una mano (to help) and the verbal syntagm are equivalent to syntagms e.g.: pagar los platos rotos (to suffer the consequences of something).</p><p>Considering those problems mentioned above and the UFs taxonomy of Zuluaga, it's proposed to develop a Diatopic Verbal Expressions Digital Dictionary (DIVEDD) for Spanish Language (diatopic subsystems of Spain and México) in order to enable the process of translation of verbal expressions (verbal locutions and verbal syntagms) in both subsystems. This prototype uses regular expressions and keywords, generating synonyms and variants expressions, finally, shows through a Corpus, examples of real use.</p><p>The paper is organized as follows. The second section describes related work with processing and translation of UFs; third section presents the architecture of the DIVEDD; the results are showed in the fourth section; and finally, conclusions are presented in the fifth section.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Work</head><p>The group of UFs constitutes an important part of the lexical system, where monolingual and bilingual dictionaries only capture certain number of units, often reduced, to an alphabetical process of selection and random description <ref type="bibr">[2]</ref>. In México there are not recent works of compilation of expressions, some of them are: the Diccionario breve de mexicanismos <ref type="bibr">[3]</ref>, the Diccionario ejemplificado de mexicanismos <ref type="bibr">[4]</ref>, and the Diccionario del español usual en México <ref type="bibr">[5]</ref>. The lack of strict rules at the time of integrate these dictionaries brought the introduction of different subsets of UFs.</p><p>There are some works related with translation of expressions in the Spanish language such as Recopilación de proverbios, proverbs which were translated into four languages (English, French, German and Italian) <ref type="bibr">[6]</ref>. In Spagnolo-Italiano: Espressioni idiomatiche e proverbi, there are a summary of idiomatic expressions, proverbs and Spanish and Italian pragmatic sentences <ref type="bibr">[7]</ref>. In <ref type="bibr">[8]</ref> there are 877 refranes españoles, sayings with their correspondence Catalan, Galician, Basque, English and French. Finally, Divergencias en la traducción de expresiones idiomáticas y refranes by Corpas Pastor <ref type="bibr">[9]</ref>, that provides a more systematic methodology for the translation of expressions between French and Spanish (Spain). This model of bibliographical record considers different uses, the level of the speaker's registration, antonyms, synonyms, source of the expression and examples among other data. This work was considered as a starting point and taking the benefits of a corpus for showing use of actual situations of expressions.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Translation of Verbal Expressions and Context of Use Extraction Through a Corpus on</head><p>Web 3</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Prototype Architecture</head><p>The architecture proposed of the DIVEDD is organized in three modules: the database, that contains the essential characteristics of the expression; the corpus, that contains in this first stage of digital texts and transcribed oral language; and the expansion expressions module, that is complemented with a list of stop-words and a database storing verbal conjugations. In the fig. <ref type="figure" target="#fig_44">1the</ref> architecture of the DIVEDD is shown. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Variants, Synonyms and +Frequents Expressions</head><p>Variants: those expressions that vary or omit any of its closed lexical elements without having semantic change. Synonyms: those expressions that have changed in their non closed lexical element i.e. key-word or those which do not contain any element in common, but they do not have a semantic change. +Frequent: the most used or most likely expression to appear in the dictionary, within a set of variations. Thus, +Frecuent expression is taken as representative.</p><p>Table <ref type="table" target="#tab_8">1</ref> shows an example of synonym expressions, therefore, the expressions: ir al bote, ir al tambo, ir a la sombra and ir tras las rajas are synonyms, because their keywords (kw) changes but they have the same definition. The same situation applies to the expressions hacer la barba, hacer la pelota and hacer la rosca, but in addition, hacer la barba is the translation into Spanish (México) of the expressions of Spain hacer la pelota and hacer la rosca.  Table <ref type="table" target="#tab_10">2</ref> shows variants through regular expressions. A regular expression is a set of pattern matching rules encoded in a string according to certain syntax rules <ref type="bibr">[10]</ref>. Thus, it is possible to describe or represent a set of strings without need to enumerate all of its elements. The operators used in the right column of Table <ref type="table" target="#tab_10">2</ref>   </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>D. B. fixed expression.</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Expansion of expressions</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">The Database</head><p>This module is based on a relational model that provides mechanisms that guarantee to avoid duplicity of records and inconsistency problems; it also guarantees the referential integrity and favors improvements of processing of the expressions. Table <ref type="table" target="#tab_14">3</ref> shows the most important attributes of verbal expressions to store.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Translation of Verbal Expressions and Context of Use Extraction Through a Corpus on Web 5</head><p>Table <ref type="table" target="#tab_14">3</ref>. More important attributes of the verbal expressions.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>ATTRIBUTE DESCRPTION</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Verb</head><p>Main verb used in the expression.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Canonical_ Verbal_ Expression</head><p>Locution or verbal syntagm in its canonical form.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Definition</head><p>Definition of the verbal expression. Field used to make the translation among the diatopic verbal expressions.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Source</head><p>Resource where the expression was extracted.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Use_ Frecuency</head><p>The number of frequencies of appearance of the expression in the corpus. Linguistic_ Record Level of registration of the expression. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Country</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Corpus Processing for DIVEDD</head><p>The corpus of the DIVEDD is conformed by written language and transcribed oral language <ref type="bibr">[11]</ref>, based on the recommendations of Sinclair J. <ref type="bibr">[12]</ref>, <ref type="bibr">[13]</ref>. The subcorpus of written language is built from digital Mexican newspapers (four sections: local news, police section, opinion section and shows section over geographical limitation in the states of D.F., Mexico, Hidalgo, Morelos, Puebla and Tlaxcala). The subcorpus of transcribed oral language is conformed by the Sociolinguistic Corpus of the México City (CSCM) <ref type="bibr">[14]</ref>. Fig. <ref type="figure" target="#fig_1">2</ref> shows the processing for the last one subcorpus. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4">Extraction Expressions Module</head><p>The module expressions expansion serves as a liaison between the DB and the corpus of DIDEVD. This module has two contributions: generate synonyms and variations of the +Frequent expression stored in the DB; and extract examples of actual usage throw the corpus.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Removing punctuation marks</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Removing non-alphabetic signs</head><p>Removing labels phonic-acoustics - <ref type="figure">-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------</ref>6 Arturo Velasco, María J. Somodevilla, and Ivo. H. Pineda Generation of Synonyms and Variants. The generation of synonyms and variants of a +frequent expression requires the entry of the possible combinations that can occur between kw´s and connectors-words or stop-words (sw). To describe all these expressions without the need to enumerate each one of them, the regular expressions are used. The operators are shown in Table <ref type="table">4</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 4. Operators of the regular expressions.</head><p>Operator Function</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>'[' y ']'</head><p>Denotes the set of verbs that can be used in the expression.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>'(' y ')'</head><p>Denotes the set of connector-words between the verb and key-word.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>'{' y '}'</head><p>Denotes the set of key-words that are used in the expression.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>','</head><p>Separator of a set of words (verbs, key-words, connector-words). Can be use ',' instead of '|'.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>'|'</head><p>Performs the same function as ','.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>'_'</head><p>Joint two or more nonseparable words in an expression</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>' '</head><p>The blank space denotes the separation between groups.</p><p>Considering the operators used in regular expressions specified in Table <ref type="table">4</ref>, the canonical verbal expression formed by hablar más que un loro, where the kw is loro and the regular expressions are denoted by:</p><p>[</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>hablar,platicar] (más_que_un,como,como_un) {loro,perico,merolico} [hablar,platicar] (más_que_una,como_una) {cotorra}</head><p>The set of variants of hablar más que un loro are: hablar como loro, hablar como un loro, platicar más que un loro, platicar como loro and platicar como un loro.</p><p>The set of all its synonyms are: hablar más que un perico, hablar como perico, hablar como un perico, platicar más que un merolico, platicar como merolico, platicar como un merolico, hablar más que una cotorra, hablar como una cotorra, platicar más que una cotorra and platicar como una cotorra.</p><p>As it can be seen the properties of regular expressions help us to match any possible variation of the expressions in the corpus without necessity of having enumerated each one.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Extraction of Usage Examples. The second contribution of extraction expressions module is the search for examples of real use of expressions stored in the DB.</head><p>The search can be performed by matching between the expression and a fragment of the corpus. The second way is to find expression through similarity. Based on the premise that the kw is crucial in the expression, The first step is to identify all the words that have a high degree of similarity with the kw was carried out. The similarity function between two strings is described in <ref type="bibr">[15]</ref> and showed below.</p><p>The second step is to identify words that preceding to the kw. Only verbs and sw´s are accepted. Another word unidentified will provoke that the fragment is rejected. The process consists of a retreat from the position of the kw. Ends in a satisfactory manner when encountering a verb and sw`s or if two verbs (an auxiliary and non auxiliary) and sw´s are matched.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Translation of Verbal Expressions and Context of Use Extraction Through a Corpus on</head><p>Web 7</p><p>//Similarity of two strings; return the percentage function similarity($s1, $s2) { $m = strlen($s1); $n = strlen($s2); $matrix = array(array($m),array</p><formula xml:id="formula_2">($n)); for($i=1; $i &lt; $m; $i++) $matrix[$i][0]=0; for($j=0; $j &lt; $n; $j++) $matrix[0][$j]=0; for ($i=1; $i &lt;= $m; $i++) { for ($j=1; $j &lt;= $n; $j++) { if ($s1[$i-1]==$s2[$j-1]) $matrix[$i][$j] = $matrix[$i-1][$j-1] + 1; else if ($matrix[$i-1][$j] &gt;= $matrix[$i][$j-1]) $matrix[$i][$j] = $matrix[$i-1][$j]; else $matrix[$i][$j] = $matrix[$i][$j-1]; } } $avgs = ($m + $n) / 2; return ($matrix[$m][$n] / $avgs) * 100; }</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Results</head><p>The extraction process using regular expressions have shown little recovery since it requires a tie with the exact expressions given by the lexicographer. On another hand, the extraction process by similarity functions between kw´s and words in the corpus heralds not only an examples extraction process; also, variant expressions can be extracted to perform in a future work, the reverse process, i.e. creating regular expressions.</p><p>Thus, besides of the similarity applied and described in <ref type="bibr">[15]</ref>, the Levenshtein similarity was applied. This function calculates the minimum number of operations (insertion, deletion or substitution) required to transform one string into another. The results are not as regulars as <ref type="bibr">[15]</ref>, and has trouble distinguishing different kw´s, and grouping kw´s of which varies only in gender and number.</p><p>The DIVEDD database was developed in MySQL 5.0.18. All the processing is implemented in PHP 5.1.2. The fig. <ref type="figure" target="#fig_2">3</ref> shows the DIVEDD interface.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Conclusions and ongoing work</head><p>The DIVEDD appears to be a system for human translation assisted by computer, providing a definition and basic characteristics of verbal expressions. The DIVEDD does not try to be a detailed dictionary but it is as a mechanism reliable of storage of</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Translation of Verbal Expressions and Context of Use Extraction Through a Corpus on</head><p>Web 9</p><p>phraseological information enforcing structure and integrity of data, reducing times of search and translation. On the other side, the mechanisms of search expressions by expressions´ attributes and its combinations make the DIVEDD a flexible tool.</p><p>Finally, note that the extraction process starting from similarity on kw´s showed more encouraging results, but with noise (information not relevant to the query), because there are parts of texts recovered, that are not relevance to the phraseology, i.e. there are non-verbal expressions. We will work on linguistic heuristics to reduce the noise.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Dynamic Concept-Based Taxonomy used for image recovery based on their textual description</head><p>Jaime Lara, María de la Concepción Pérez de Celis, David Pinto Facultad de Ciencias de la Computación, Benemérita Universidad Autónoma de Puebla, Puebla, México. {jlara, cperezdecelis, dpinto}@cs.buap.mx</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>An alternative to classical search engines (keyword-based search) is to allow the user to navigate through contents. This process of navigation requires the user to know the way in which the information is organized. A practical solution to this inconvenient is the use of a pre-established taxonomy. We must point out that taxonomies require the organization of contents around a certain knowledge domain. From our point of view, a taxonomy improves the recovery of information on specific topics, but it is still a rigid solution since it does not allow for the possibility of restructuring the users searches as he advances in the recovery of results. As such, the user cannot connect the recovered objects with other knowledge domains within such objects may be connected.</p><p>A possible solution that allows the user to be guided and also by self may interact and refine his search is the usage of what is known as a faceted taxonomy. Such taxonomy allows the information to be structured and accessed in more than one dimension. This benefits the user because he can easily and intuitively localize and explore the information using the different approaches provided by the taxonomy. In this work we will focus on the methodology used to build a taxonomy which will be the base for the implementation of a system of faceted classification, which will then allow the user to build his own knowledge map. In the following sections we will present some of the work linked to the design and construction of taxonomies and we will analyze its usage in the recovery of information. Later on, we will discuss the methodology used for the confirmation of the taxonomy, with emphasis on the recovery of objects based on their textual descriptions. We will present the results we have obtained and we will emphasize the importance of conceptualizing the operations that allow for the creation of open taxonomies.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Related Works</head><p>An information retrieval system in which one has a collection of objects that have diverse properties, the relevance of such properties varies depending on the user. As a matter of fact, a solution for finding information regarding a particular topic consists in asking an expert on the subject. Due to the fact that, generally, this is not possible, one has to recur to the use of taxonomies.</p><p>The importance of taxonomies lies in the possibility of them being used as a triplet (classification scheme, semantic interpretation, knowledge map <ref type="bibr">[1]</ref>) in the organization and recovery of collected objects (possibly in a database).</p><p>Taxonomies can be organized under different structures: lists, hierarchies, polihierarchies, multi-dimensional matrix and facets, among others. Among them, polihierarchies, multi-dimensional matrix and facets are linked to the possibilities of objects belonging to more than one category.</p><p>Characterizing objects under diverse categories or characteristics allows us to create a metadata model were the objects and their characteristics take on a new meaning. It is because of such a phenomenon that faceted taxonomies can take advantage of the way in which metadata behave. Metadata models are a collection of information built on some type of object, or on a part of such an object. For example, the name (or title) of an artwork, its author, date, image, textual description (denotation), interpretation (connotation) or genre, represent metadata that can be associated to a cultural object.</p><p>As such, every element in the metadata scheme can be incorporated as a concept of the faceted taxonomy, and thus it can be recovered using a search engine. Therefore it is possible to access an object under any of the dimensions under which it has been classified. If we consider the previous example, we could recover a cultural object by its genre, connotation or denotation or we could navigate through the different facets, taking into account that they are orthogonal among themselves.</p><p>Sacco <ref type="bibr">[2]</ref>, <ref type="bibr">[3]</ref> introduced the concept of dynamic taxonomies and the notion that it could withstand the incorporation of facets that, in themselves, require an independent taxonomy for their description. Due to the fact that our objective is the recovery of images based on their textual description, we will use the fundamentals of dynamic taxonomies but we will also extend the domain of the data modeling created by Sacco, because we will include textual objects.</p><p>In the literature we can find two interesting approaches for the management of textual objects. The first such approach assumes the existence of various topics over which one wishes to generate taxonomy, and because of that the algorithms are designed to extract the terms and concepts linked to such topics from the documents. The MindMap system <ref type="bibr">[4]</ref>, in particular, proposes the generation of multiple taxonomies for any collection of documents, each one with unique topics. This multiple taxonomies are visualized in the system as an integrated tool, thus obtaining a system that allows the organization of information in multiple ways. In this system, the classification of a document under any taxonomy depends on its similarities with other documents. For this purpose, a system of spatial coordinates is used, in which the similarities are determined by the proximity between the coordinates of each document. As we have pointed out before, in this approach each one of the multiple taxonomies requires a series of keywords, associated to the concepts under which the classification of the documents will be structured, in order for the analysis to begin.</p><p>In contrast, the second approach is centered in determining the different facets and defining their taxonomy from the textual analysis of the collected documents by using text analysis. A good example is the algorithm used by the Flamenco project <ref type="bibr">[5]</ref> of Berkley University for the automatic generation of facets using WordNet over a corpus written in English.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Our approach</head><p>During the development of the information system for the management of cultural objects, we have implemented the metadata model proposed by CCO <ref type="bibr">[6]</ref>. However, when extending this model with the metadata of genre, connotation and denotation, the inclusion of a textual description of the objects was necessary. This fact determines the necessity of a methodology that allows for the generation of a taxonomic structure of the facets, under which the different terms included in the description of the objects will be classified. We must point out that such a methodology can be extended to any corpus of descriptive documents over which one means to develop a taxonomy. Our proposal consists in using techniques and algorithms employed in information retrieval to generate a faceted classification, which allows for the association of part of a document to the different facets. Such facets will be generated from a thesaurus linked with a "controlled vocabulary". This vocabulary, in turn, will be extracted fro the processing of different texts from the target collection. This process is described in detail in the following sections.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Approach by dynamic taxonomy</head><p>The largest difference between a conventional taxonomy and a dynamic taxonomy is that the former are monodimensional (one element is classified under one concept), while the former are multidimensional (Fig. <ref type="figure" target="#fig_0">1</ref>). The term dynamic reflects the ability of the taxonomy to adapt to different approaches, perspectives and interests.</p><p>Sacco <ref type="bibr">[2]</ref>, <ref type="bibr">[3]</ref> established the following inference rule: Two concepts A and B are related iff there is at least one item D in the infobase which is classified at the same time under A (or under one of A's descendants) and under B (or under one of B's descendants). The left side of Fig. <ref type="figure" target="#fig_1">2</ref> shows a set of data classified under two facets, and in the right side can be appreciated the change that the taxonomies of our facets go through when we focus our search on the concepts B and H. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Study case</head><p>The data set that will be studied consists of a total of 500 artworks <ref type="bibr">[7]</ref>, <ref type="bibr">[8]</ref> (An example can be appreciated on Fig. <ref type="figure" target="#fig_2">3</ref>) with description in the Spanish language, of which a training set of 200 artworks was selected for testing.</p><p>In order to implement the dynamic taxonomy we used the following facets: genre, connotation and denotation. Genre refers to each one of the different categories or classes under which the artworks can be classified according to common forms and contents. For example: portraits, self-portraits, landscapes, religion, mythology etc. Of these three facets, only genre is defined by an expert user, while the other two are obtained automatically. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Methodology</head><p>In the following paragraphs we will show the steps taken in order to generate a classification system based on dynamic taxonomies.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1">Obtaining a Controlled Vocabulary</head><p>A Controlled Vocabulary (CV) is an organized lists of words and phrases that are used to initially tag content, and then to find it through navigation or search.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.1">Elimination of stopwords</head><p>The first step consists in the elimination of stopwords, since such words do not allow for discrimination of relevant attributes of the objects. There are several lists of stopwords in Spanish (Snowball<ref type="foot" target="#foot_5">1</ref> ; Ranks<ref type="foot" target="#foot_6">2</ref> ).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.2">Stemmer</head><p>In order to obtain a dynamic taxonomy, one must have a controlled vocabulary. In this particular case we applied a Spanish stemmer<ref type="foot" target="#foot_7">3</ref> based on Porter's algorithm, which allows for the decrement of the controlled vocabulary and augments the definition of each concept, providing a better recall.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.3">N-Grams</head><p>With the objective of finding concepts formed by more than a word, we obtained bigrams, trigrams and 4-grams of the textual description of artworks. And also, we have a set of words (unigrams).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2">Defining the Concepts</head><p>A concept is a label which identifies a set of documents <ref type="foot" target="#foot_8">4</ref> (classified under that concept), every concept is related with a certain level of abstraction that depends with the level in the taxonomy, this make clear that concepts are not terms. So, the problem here is how we can know what terms of phrases of our controlled vocabulary can be a concept, to answer that question we use a thesaurus, using a thesaurus we can define the part of our controlled vocabulary that we can use as concepts as we can see in the Fig. <ref type="figure" target="#fig_3">4</ref> and in the equation 1.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>CV Thesaurus Concepts</head><p>.</p><p>(1)</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Fig. 4. A Concept definition</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.1">Incrementing and Expanding Concepts</head><p>By using clustering algorithms, we look for terms that were not originally considered part of the concepts using the remaining concepts of the thesaurus. Afterwards, new concepts are added using a supervised process like the one shown in Fig. <ref type="figure" target="#fig_4">5</ref>. Also, there is a possibility to expand the definition of each concept, this leads us to the generation of new clusters between the controlled vocabulary and the concepts. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3">Obtaining taxonomies</head><p>The next step consists of generating the taxonomies of the concepts using the thesaurus hierarchies.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.1">Defining the Taxonomy</head><p>In this step we obtain the hierarchical structure of every concept based on the thesaurus structure, we use a hash table to do this process (See Fig. <ref type="figure" target="#fig_24">6</ref>). The taxonomy is then created by the process of linking all the concepts. Fig. <ref type="figure" target="#fig_24">6</ref>. Obtaining the hierarchy of every concept.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.2">Structuring the Facets</head><p>Once the hierarchic structure of the concepts contained in the vocabulary is ready, one must supervise under which facet they will be placed. In  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.3">Pruning</head><p>The taxonomy should be pruning, to avoid that the user spend time expanding unnecessary nodes, this nodes are those that doesn't helps us in the filter process.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.4">Frequency filter</head><p>If we order the controlled vocabulary based on its frequency, it is possible to eliminate the less frequent concepts, due to the fact that they will generally not be used for information recovery. However, our proposal consists in implementing this filter directly over the taxonomy. This process entails a second pruning over the faceted taxonomies.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.4">Indexing</head><p>Each artwork contains a textual description. Such description has words or phrases that are included in the controlled vocabulary. At first, and before we have used hierarchies to bond the controlled vocabulary to the faceted taxonomy, the indexes are not related among each other (Fig. <ref type="figure" target="#fig_17">8</ref>). However, once created the taxonomies, each index is related to each other by a hierarchical scheme, as shown in Fig. <ref type="figure" target="#fig_18">9</ref>. This bonding allows to index the artwork under the information that contains its textual description, and to add the hierarchical information of each concept (Fig. <ref type="figure" target="#fig_19">10</ref>).   </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.6">Implementation (Navigation Tree)</head><p>The final step is make a visual framework in which the user can select and combine appropriate concepts <ref type="bibr">[2]</ref>, this is develop by a Navigation Tree <ref type="bibr">[9]</ref> (taxonomic tree <ref type="bibr">[1]</ref>). The navigation tree contains nodes that enable the user to start browsing in one facet and then cross to another, and so on, until reaching the desired level of specificity <ref type="bibr">[9]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Results and Evaluation</head><p>By using the procedure described earlier we obtained a controlled vocabulary of 500 concepts that describe the 200 documents. It is important to mention that the eliminated stopwords represented 50% of the total words. By means of this procedure we were able to bond each artwork to an average of fourteen indexes.</p><p>We performed a comparison between our faceted classification system and the "Full Text Search" function of MySQL, which uses a Boolean search for any given data set. We configured a set of supervised consults, taking into account two criteria: recall and precision. Fig. <ref type="figure" target="#fig_21">12</ref> shows the results.</p><p>As one can see in Fig. <ref type="figure" target="#fig_21">12</ref>, the faceted classification system provides a significantly higher degree of recall when compared to the Boolean search, underlining the advantage of using a conceptual search instead of a textual search. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.">Conclusions</head><p>So far, the automatic solutions for hierarchy constructions available have not offered satisfactory outcomes when faced with the construction of taxonomies. However, some methodologies developed for the construction of facets and the generation of taxonomies based on textual analysis has yielded encouraging results. As of now we are implementing new algorithms that will allow the generation of interpretations based on generic descriptions, expanding the "connotation" facet described in this paper. We have also considered the usage of a terminological conceptual thesaurus for this task, and the possibility of allowing the user to use "open taxonomies" thus allowing the taxonomy to evolve, reflecting the progression of the words and their interpretation and keeping its novelty.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Classifying people names on the web is a task that requires a special attention by the classification societies community. Searching people in Internet is one of the most common activities performed by the World Wide Web users <ref type="bibr">[1]</ref>. The main challenge consists in bringing together all the results (of a given search engine) that share the same occupation/profession (which very often are ambiguous) by using a highly scalable classification method.</p><p>In this paper, we report the results obtained in the Web People Search task [2] when using a system based on hash-based text retrieval techniques <ref type="bibr">[3]</ref>. In particular, we have constructed a new vectorial coordinates system for the representation of the original data and, thereafter, we calculate the distance of the vectorial representation of each input dataset by means of a hash function.</p><p>The experiments were carried out by using the WePS-2 collection. In summary, it is made up of 30 ambiguous names of people, each name with a number of html pages with information related with that people name. The complete description of the evaluated corpus is given into detail in <ref type="bibr">[2]</ref>.</p><p>We must take into account that the fingerprinting technique may allow indexing and clasifying of documents in a one single step. Therefore, given the huge amount of information available in Internet, we consider that the main contribution of this research work consists of providing a very fast way of classifying people names on the World Wide Web.</p><p>The evaluation of the experiments carried out show that the implemented technique could have a positive impact in the analysis/indexing of huge volumes of information. However, the feature set for all the documents in the WePS framework needs to be further investigated.</p><p>The remainder of this document is structured as follows. Section 2 presents the document fingerprint technique used in the process of indexing and clustering of documents in the Web People Search framework. In Section 3 we describe the components of the implemented system. The experimental results are discussed in Section 4. Finally in Section 5 the conclusions are given.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Document Fingerprinting</head><p>Document indexing based on fingerprinting is a powerful technology for similarity search in huge volumes of documents. The goal is to provide a proper hash function which cuasi-uniquely identifies each document, so that the hash collisions may be interpreted as similarity indication.</p><p>Formally, given two documents d 1 and d 2 , and the fingerprint of the two documents h(d 1 ) and h(d 2 ), respectively. We consider d 1 and d 2 to be ǫ-similar</p><formula xml:id="formula_3">iff |h(d 1 ) − h(d 2 )| &lt; ǫ.</formula><p>In the context of document indexing/clustring/retrieval a fingerprint h(d) of a document d may be considered as a set of encoded substrings taken from d, which serve to identify d uniquely.</p><p>Defining the specific hash function to encode the substrings of the documents is the main challenge of the fingerprinting technique. In particular, in the implementation of the BUAP Web People Search system we defined a small set of term-frequency vectors (which are used as reference for a new system coordinates) in order to be considered as the new reference for the vectorial representation of each document of the WePS-2 collection. In Figure <ref type="figure" target="#fig_0">1</ref> we may see an overview of the proposed approach.</p><p>Formally, given a set of k reference vectors, {r 1 , r 2 , • • • , r k }, and the vectorial representation of a document d. We defined the fingerprint of d as shown in equation <ref type="bibr">(1)</ref>.</p><formula xml:id="formula_4">h(d) = k i=1 r 1 • d<label>(1)</label></formula><p>The specific features used in the vectorial representation of the documents are explained in the following section.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">The BUAP system</head><p>The system comprises the following components: Pre-processing: We have programmed two implementations in order to perform the HTML to text conversion. The first HTML to text converter was programmed with Java, whereas the second was implemented with AWK.</p><p>No HTML tags nor url's were considered in the text extraction. Named entity recognition: We used the Stanford Named Entity Recognizer <ref type="bibr">[4]</ref> in order to extract names of places, organizations and people names from the target documents. Document representation: The features used to represent each document comprised all the named entities recognized by the Stanford NER. Thus, the vectorial representation of a document d was:</p><formula xml:id="formula_5">d = {tf (ne 1 ), tf (ne 2 ), • • • , tf (ne n )},<label>(2)</label></formula><p>where tf (ne i ) is the frequency of the i-th named entity recognized in the document d. Reference vector generation: In order to generate the k reference vectors, we calculated the named entity vocabulary, V N E , of the entire target collection. We sorted this vocabulary in a non-increased order according to the named entity frequency (over the complete collection) and, thereafter, we selected only those named entities whose frequency were between the so called "transition" range <ref type="bibr">[5]</ref>. The transition range allows to obtain the mid-frequency terms of a given vocabulary.</p><p>A typical formula used to obtain the center of the transition range (transition point) is given in Equation (3).</p><formula xml:id="formula_6">T P V = √ 8 * I 1 + 1 − 1 2<label>(3)</label></formula><p>where I 1 represents the number of terms (in our case, named entities) with frequency equal to 1.</p><p>Once the transition point has been found, we may extract the mid-frequency named entities, which are those which obtain the closest frequency values to T P V , i.e.,</p><formula xml:id="formula_7">V T P = {ne i |ne i ∈ V N E , U 1 ≤ tf c (ne i ) ≤ U 2 },<label>(4)</label></formula><p>where tf c (ne i ) is the frequency of the i-th entity over the complete document collection and U 1 is a lower threshold obtained by a given neighbourhood value of the TP: U 1 = (1 − N T P ) * T P V , where 0 ≤ N T P &lt; 1. U 2 is the upper threshold and it is calculated in a similar way: U 2 = (1+N T P ) * T P V . Thus, the representation of the j-th reference vector r j is given as follows:</p><formula xml:id="formula_8">r j = {tf c (ne 1 ), tf c (ne 2 ), • • • , tf c (ne m )}. (<label>5</label></formula><formula xml:id="formula_9">)</formula><p>Indexing/clustering: The indexing process was carried out by using the formula expressed in Equation ( <ref type="formula" target="#formula_4">1</ref>). We used a specific threshold (ǫ) in order to determine a range of hash-based values (documents) that should belong to the same cluster. The overlapping of clusters was not considered but it may be easily implemented.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Experimental results</head><p>Besides the evaluation of the WePS-2 collection, we performed a set of experiments over the training and test dataset of the WePS-1 collection (see <ref type="bibr">[1]</ref> for a complete description of these datasets). The obtained results are presented in Tables 1, 2 and 3, respectively. In these tables we may see the following set of metrics used to evaluate the performance of the implemented system: BEP: BCubed Precision BER: BCubed Recall FMeasure 0.5 BEP-BER: F-measure of B-Cubed P/R with alpha set to 0.5 FMeasure 0.2 BEP-BER: F-measure of B-Cubed P/R with alpha set to 0.2 P: Purity IP: Inverse Purity FMeasure 0.5 P-IP: F-measure of Purity and Inverse Purity with alpha set to 0.5 FMeasure 0.2 P-IP: F-measure of Purity and Inverse Purity with alpha set to 0.2</p><p>For more details about the evaluation metrics please refer to <ref type="bibr">[6]</ref>. The baselines and the rationale for F -measures with alpha 0.2 are explained in the WePS-1 task description paper <ref type="bibr">[1]</ref>.</p><p>We have tested six differents approaches varying the document similarity threshold (ǫ), the HTML to text converter and the use or not of named entities. The name of each approach as well as their describtion is given as follows:</p><p>BUAP 1 (ǫ = 0.0004): Java-based HTML to text converter without NER, i.e., all the document terms are used. BUAP 2 (ǫ = 0.0004): Java-based HTML to text converter with NER, i.e., all the document named entities are used. BUAP 3 (ǫ = 0.0004): AWK-based HTML to text converter without NER, i.e., all the document terms are used. BUAP 4 (ǫ = 0.0004): AWK-based HTML to text converter with NER, i.e., all the document named entities are used. BUAP 5 (ǫ = 0.3): Java-based HTML to text converter without NER, i.e., all the document terms are used. BUAP 6 (ǫ = 0.3): AWK-based HTML to text converter without NER, i.e., all the document terms are used.</p><p>We may see (Tables 1, 2 and 3) that the implemented approaches obtained a performance comparable with two of the proposed baselines, ALL IN ONE and ONE IN ONE, with a document similarity threshold (ǫ) equal to 0.3 and 0.0004, respectively. The selection of any of the two HTML to text converters was not important with respect to the obtained results. Moreover, we could not confirm the benefit of using named entities with respect to those approaches that did not use them, because the obtained results did not show significant difference among the different approaches.</p><p>Although, some of the implemented approaches obtained acceptable results in comparison with the baselines, in the case of the WePS-2 collection any of the first four approaches outperformed the proposed baselines. We consider that the expected document distribution over the final clusters has played an important role on the obtained results, since the presented algorithm of fingerprinting usually assumes a uniform distribution of documents over the discovered clusters.</p><p>The evaluation of the experiments carried out show that the implemented technique could have a positive impact in the analysis/indexing of huge volumes of information. However, the feature set for all the documents in the WePS framework needs to be further investigated.</p><p>As future work, we would like to experiment on feature selection in order to clearly benefit the construction of the reference vector set. Although we would like to keep the process as unsupervised as possible, we are considering the use of supervised classifiers in order to extract the most important features to tackle the Web People Search task.</p><p>Finally, we would like to analyse the use new hash-based functions and new document representations which consider characteristics other than only term or named entity frequencies.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Conclusions</head><p>We implemented a hash-based function in order to uniquely identify each document from a text collection in the framework of the Web People Search task. The hash collisions were interpreted as similarity degree among the target documents. In this way, we constructed an algorithm which only takes into account the local features of each document in order to index/cluster them. The experimental results over the WePS-1 test and training datasets showed an acceptable performance of the proposed algorithm. However, the proposed reference vector for the fingerprinting-based model were useless when evaluating with the WePS-2 dataset. The proper construction of reference vectors for the automatic and unsupervised classification of people names in the Web needs to be further investigated.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>mQA: Question Answering in Mobile devices</head><p>Fernando Zacar ías F. 1 , Alberto Tellez V. 2 , Marco Antonio Balderas 3 , and Rosalba Cuapa C. 4   Benemérita Universidad Autónoma de Puebla, 1,3,4 Computer Science and 2 Collaborator -INAOE 14 Sur y Av. San Claudio, Puebla, Pue. 72000 México 1 fzflores@yahoo.com.mx, 2 albertotellezv@ccc.inaoep.mx 3 balderasespmarco@gmail.com, 4 rcuapa canto@yahoo.com Abstract. In this paper, we present a novel proposal for Question Answering through mobile devices. Thus, an architecture for a mobile Question Answering system based on WAP technologies is deployed. The architecture propose moves the issue of Question Answering to the context of mobility. This paradigm ensures that QA is seen as an activity that provides entertainment and excitement pleasure. This characteristic gives to QA an added value. Furthermore, the method for answering definition questions is very precise. It could answer almost 90% of the questions; moreover, it never replies wrong or unsupported answers. Considering that the mobile-phone has had a boom in the last years and that a lot of people already have mobile telephones (approximately 3.5 billions), we propose an architecture for a new mobile system that makes QA something natural and effective for work in all fields of development. This obeys to that the new mobile technology can help us to achieve our perspectives of growth. This system provides to user with a permanent communication in anytime, anywhere and any device (PDA's, cell-phone, NDS, etc.).</p><p>Keywords: Mobile devices, Question Answering, WAP, GPRS.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Each generation of mobile communications has been based on a dominant technology, which has significantly improved spectrum capacity. Until the advent of IMT-2000, cellular networks had been developed under a number of proprietary, regional and national standards, creating a fragmented market.</p><p>-First Generation was characterized for Advanced Mobile Phone System (AM-PS). It is an analog system based on FDMA (Frequency Division Multiple Access) technology. However, there were also a number of other proprietary systems, rarely sold outside the home country. -Second Generation, it includes five types of cellular systems mainly:</p><p>• Global System for Mobile Communications (GSM) was the first commercially operated digital cellular system. • GSM uses TDMA (Time Division Multiple Access) technology.</p><p>• TDMA IS-136 is the digital enhancement of the analog AMPS technology. It was called D-AMPS when it was fist introduced in late 1991 and its main objective was to protect the substantial investment that service providers had bmade in AMPS technology. • CDMA IS-95 increases capacity by using the entire radio band with each using a unique code (CDMA or Code Division Multiple Access) • Personal Digital Cellular (PDC) is the second largest digital mobile standard although it is exclusively used in Japan where it was introduced in 1994.</p><p>• Personal Handyphone System (PHS) is a digital system used in Japan, -Third Generation, better known as 3G or 3rd Generation, is a family of standards for wireless communications defined by the International Telecommunication Union, which includes GSM EDGE, UMTS, and CDMA2000 as well as DECT and WiMAX. Services include wide-area wireless voice telephone, video calls, and wireless data, all in a mobile environment. Thus, 3G networks enable network operators to offer users a wider range of more advanced services while achieving greater network capacity through improved spectral efficiency.</p><p>Currently, mobile devices are part of our everyday environment and consequently part of our daily landscape <ref type="bibr">[5]</ref>. The current mobile trends in several application areas have demonstrated that training and learning no longer needs to be classroom. Current trends suggest that the following three areas are likely to lead the mobile movement: m-application, e-application and u-application. There are estimated to be 2.5 billion mobile phones in the world today. This means that this is more than four times the number of personal computers (PCs), and today's most sophisticated phones have the processing power of a mid-1990s PC. Even, in a special way, many companies, organizations, people and educators are already using iPhone, iPod, NDS, etc., in their tasks and curriculas with great results. They are integrating audio and video content including speeches, interviews, artwork, music, and photos to bring lessons to life. Many current developments, just as ours <ref type="bibr">[5,</ref><ref type="bibr">3,</ref><ref type="bibr">6]</ref>, incorporate multimedia applications.</p><p>In the late 1980's, a researcher at Xerox PARC named Mark Weiser <ref type="bibr">[4]</ref>, coined the term "Ubiquitous Computing". It refers to the process of seamlessly integrating computers into the physical world. Ubiquitous computing includes computer technology found in microprocessors, mobile phones, digital cameras and other devices. All of which add new and exciting dimensions to applications.</p><p>As pragmatic uses grow for cellphones, mobile technology is also expanding into creative territory. New public space art projects are using cellphones and other mobile devices to explore new ways of communicating while giving everyday people the chance to share some insights about real world locations.</p><p>While your cellphone now allows you to play games, check your e-mail, send text messages, take pictures, and oh, yeah, make phone calls, it can perhaps serve a more enriching purpose. Thus, we think that widespread internet access and collaboration technologies are allowing businesses of all sizes to mobilise their workforce. Such innovations provide additional flexibility without the need to invest in expensive and complex on-premise infrastructure requirements. Furthermore, it makes "eminent sense" to fully utilise the web commuting options provided by mobile technology.</p><p>The problem of answering questions has been recognized and partially tacled since the 70's for specific domains. However, with the advent of browsers working with billions of documents in internet, the need has newly emerged, having led to approaches for open-domain QA. Some examples of such approaches are emergent question answering engines such as answers.com, ask.com, or additional services in traditional nrowsers, such as Yahoo.</p><p>Recent research in QA has been mainly fostered by the TREC and CLEF conferences. The first one focus on English QA, whereas the second evaluates QA systems for most European languages except English. To do, both evaluation conferences have considered only a very restriced version of the general QA problem. They basically contemplate simple questions which assume a definite answer typified by a named entity or noun phrase, such as factoid questions (for instance, "How old is Cher?" or "Where is the Taj Mahal?") or definition questions ("Who is Nelson Mandela?" or "What is the quinoa?"), and exclude complex questions such as procedural or epaculative ones.</p><p>Our paper is structured as follows: In section 2 we describe the state of the art about QA and similar works. Next, in section 4 we present our mobile architecture to support question answering. Section 5 contains our perspectives about our future work. This work consist in incorporate answering definitions questions. Finally, the conclusions are drawn in section 6.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">The state of the art</head><p>One of the oldest problems of human history is raising questions about several issues and conflicts that torments our existence. Since children this is the mechanism we use to understand and adapt to our environment. The counterpart to ask questions is to answer the questions that we do, an activity that also requires intelligence. This activity has a difficulty level that has tried to delegate to computers, almost since the emergence of these. The issue of question answering for a computer has been recognized and tackled from the decade of the 70s century past for specific domains. In Mexico, have been obtained excellent results in this context, for this reason we propose to bring these same results with mobile technologies. Recent research has focused on developing systems for question answering to open domain, ie systems that takes as their source of information a collection of texts on a variety of topics, and solve questions whose answers can be obtained from the collection of departure. From question answering systems developed so far, we can identify three main phases:</p><p>1. Analysis of the question. This first phase will identify the type of response expected from the given question, that is expected to be a question of "when" a kind of response time, or a question "where" will lead us to identify a place.</p><p>Response rates are most commonly used personal name, name organization, number, date and place. 2. Recovery of the document. In the second stage performs a recovery process on the collection of documents using the question, which is to identify documents on the question that probably contain the kind of response expected. The result of this second stage is a reduced set of documents and preferably specific paragraphs. 3. Extraction of the response. The last phase uses the set of documents obtained in the previous phase and the expected type of response identified in the first phase, to locate the desired response.</p><p>Questions of definition require a more complex process in the third stage, since they must obtain additional information segments and at the same time are not repetitive. To achieve a good "definition" must often resort to various documents <ref type="bibr">[1]</ref>.</p><p>Currently the question answering on mobile devices for open domains is in a development stage. The project QALL-ME, is a project of 36 months, funded by the European Union and will be conducted by a consortium of seven institutions, including four academic and three industrial companies. The aim is to establish a shared infrastructure for developing a QA infrastructure via mobile phone for any tourist or citizen can instantly access to different information regarding the services sector, be it a movie in the cinema, a theater or restaurant of a certain type of food. All this in a multilingual and multimodal mode for mobile devices. The project will experiment with the potential of open domain QA and evaluation in the context of seeking information from mobile devices, a multimodal scenery which includes natural speech as input, and the integration of textual answers, maps, pictures and short videos as output.</p><p>The architecture proposed in the QALL-ME project is a distributed architecture in which all modules are implemented as Web services using standard language for defining services. In figure <ref type="figure" target="#fig_0">1</ref> shows the main modules of this architecture. The architecture of the QALL-ME described as follows:</p><p>"The central planner is responsible for interpreting multilingual queries. This module receives the query as input, processes the question in the language in which it develops and, according to the parameters of context, directs the search for required information. Extractor to a local response. The extraction of the Fig. <ref type="figure" target="#fig_0">1</ref>. Main QALL-ME Architecture <ref type="bibr">[8]</ref> response is made on different semantic representations of the information depends on the type of the original source data from which we get the answer (if the source is plain text, the semantic representation is an annotated XML document if the source is a website, the semantic representation is a database built by a wrapper). Finally, the responses are returned to the central planners to determine the best way to represent the requested information" <ref type="bibr">[8]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Movile Question Answering for Definitions Questions</head><p>The method for answering definition questions uses Wikipedia <ref type="bibr">[10]</ref> as target document collection. It takes advantage of two known facts: <ref type="bibr">[10]</ref> Wikipedia organizes information by topics, that is, each document concerns one single subject and, <ref type="bibr">[11]</ref> the first paragraph of each document tend to contain a short description of the topic at hand. This way, it simply retrieves the document(s) describing the target term of the question and then returns some part of its initial paragraph as answer.</p><p>Figure <ref type="figure" target="#fig_1">2</ref> shows the general process for answering definition questions. It consists of three main modules: target term extraction, document retrieval and answer extraction.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Finding Relevant Documents</head><p>In order to search in Wikipedia for the most relevant document to the given question, it is necessary to firstly recognize the target term. For this purpose the method uses a set of manually constructed regular expressions such as: "What-Which-Who-How"+"any form of verb to be"+&lt;TARGET&gt;+"?", "What is a &lt;TARGET&gt; used for?", "What is the purpose of &lt;TARGET&gt;?", "What does &lt;TARGET&gt; do?", etc. Then, the extracted target term is compared against all document names and the document having the greatest similarity is recovered and delivered to the answer extraction module. It is important to mention that, in order to favor the retrieval recall, we decided using the document names instead of the document titles since they also indicate their subject but normally they are more general (i.e., titles tend to be a subset of document names). In particular, the system uses the Lucene <ref type="bibr">[11]</ref> information retrieval system for both indexing and searching.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Extracting the Target Definition</head><p>As we previously mentioned, most Wikipedia's documents tend to contain a brief description of its topic in the first paragraph. Based on this fact, this method for answer extraction is defined as follows:</p><p>-Consider the first sentence of the retrieved document as the target definition (the answer).</p><p>-Eliminate all text between parenthesis (the goal is to eliminate comments and less important information). -If the constructed answer is shorter than a given specified threshold2, then aggregate as many sentences of the first paragraph as necessary to obtain an answer of the desire size.</p><p>For instance, the answer for the question "Who was Hermann Emil Fischer?" (refer to Figure <ref type="figure" target="#fig_1">2</ref>) was extracted from the first paragraph of the document "Hermann Emil Fischer": "Hermann Emil Fischer (October 9, 1852 -July 15, 1919) was a German chemist and recipient of the Nobel Prize for Chemistry in 1902. Emil Fischer was born in Euskirchen, near Cologne, the son of a businessman. After graduating he wished to study natural sciences, but his father compelled him to work in the family business until determining that his son was unsuitable".</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Evaluation Results of our method</head><p>This section presents the experimental results about the participation <ref type="bibr">[7]</ref> at the monolingual Spanish QA track at CLEF 2007. This evaluation exercise considers two basic types of questions, definition and factoid. However, this year there were also included some groups of related questions. From the given set of 200 test question, our QA system treated 34 as definition questions and 166 as factoid. Table <ref type="table" target="#tab_14">3</ref>.3 details our general accuracy results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 1. System's general evaluation</head><p>It is very interesting to notice that our method for answering definition questions is very precise. It could answer almost 90% of the questions; moreover, it never replies wrong or unsupported answers. This result evidenced that Wikipedia has some inherent structure, and that our method could effectively take advantage of it. <ref type="bibr">[7]</ref> 8 F. Zacar ías et al.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Proposed Architecture</head><p>New technologies such as Wireless Application Protocol (WAP), General Packet Radio Service (GPRS) and 3G (3rd generation) further increase communication technologies and access to information. Wireless Application Protocol (commonly referred as WAP) is an open international standard for application layer network communications in a wireless communication environment. Its main use is to enable access to the Mobile Web from a mobile phone or PDA. The recent advances in mobile technology and wireless basic requirements of the applications of WAP has helped improve trade and mobile services. This protocol is a secure specification allowing users to access services and information instantly through wireless mobile devices. WAP is composed of the following form: uses Wireless Markup Language (WML), which includes the Handheld Device Markup Language (HDML). WML can also trace its roots to eXtensible Markup Language (XML). The best known markup language is Hypertext Markup Language (HTML). Unlike HTML, WML is considered a Meta language. WAP also allows the use of standard Internet protocols such as UDP, IP and XML. Although WAP supports HTML and XML, the WML language (an XML application) is specifically devised for small screens and one-hand navigation without a keyboard.</p><p>WAP also supports WML Script. It is similar to JavaScript, but makes minimal demands on memory and CPU power because it does not contain many of the unnecessary functions found in other scripting languages. The WAP programming model is similar to the Web programming model with matching extensions, but it accommodates the characteristics of the wireless environment. Figure <ref type="figure" target="#fig_2">3</ref> illustrates this model. As you can see, the WAP programming model is based heavily on the Web programming model. But how does the WAP gateway work with HTML? In some cases, the data services or content located on the Web server is HTMLbased. Some WAP gateways could be made to convert HTML pages into a format that can be displayed on wireless devices. But because HTML wasn't really designed for small screens, the WAP protocol defines its own markup language, the Wireless Markup Language (WML), which adheres to the XML standard and is designed to enable powerful applications within the constraints of handheld devices. In most cases, the actual application or other content located on the Web server will be native WAP created with WML or generated dynamically using Java servlets or JSP.</p><p>In HTML, there are no functions to check the validity of user input or to generate messages and dialog boxes locally. To overcome this limitation, JavaScript was developed. Similarly, to overcome the same restrictions in WML, a new scripting language known as WMLScript has been developed. I'll cover more on WML and WMLScript in later sections. However, General packet radio service (GPRS) is a packet oriented mobile data service available to users of the 2G cellular communication systems, global system for mobile communications (GSM) and provides data rates of 56-114 kbit/s.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>GPRS upgrades GSM data services providing:</head><p>-Multimedia messaging service (MMS) -Short message service (SMS) -Push to talk over cellular (PoC/PTT) -Instant messaging and presence-wireless village -Internet applications for smart devices through wireless application protocol (WAP) -Point-to-point (P2P) service: inter-networking with the Internet (IP) Some mobile phone operators offer flat rate access to the Internet, while others charge based on data transferred, usually rounded up to 100 kilobytes. During the heyday of GPRS in the developed countries, around 2005, typical prices varied from EUR C 0.24 per megabyte to over C 20 per megabyte. In developing countries, prices vary widely, and change. Some operators gave free access while they decided pricing, for example in Togocel.tg in Togo, West Africa, others were over-priced, such as Tigo of Ghana at one US dollar per megabyte or Indonesia at $3 per megabyte. AirTel of India charges $0.025 per megabyte, and Telstra of Australia charges $22.53 per megabyte. As of 2008, data access in Canada is still prohibitively expensive. For example, Fido charges $0.05 per kilobyte, or roughly $50 per megabyte. In Venezuela, Digitel charges about $20 per 100 Mb or $25 for unlimited access. In Mexico charges $.04 per Kb. or roughly $40 per megabyte.</p><p>This WAP and GPRS infrastructure gives us the appropriate stage to propose the following architecture for QA on mobile devices.  In Figure <ref type="figure" target="#fig_3">4</ref> shows the proposed architecture for the definition question answering (DQA), the module will be implemented question answering as Web services using standard languages. It takes a natural language query from a mobile device, this question is sent to the web service which returns a specific answer from a collection of information sources.</p><p>The proposed architecture consists of the following modules:</p><p>-Mobile Question Answering (mQA): is the interface of the mobile device to the user, which is responsible for communicating via GPRS to the web service DQA. -Web Service Definition Question Answering (DQA): is the web service to meet all the demands of mQA and from the site, to send all questions to LabTL QA engine to be answered -"LabTL QA engine: is responsible for seeking the answer to the definition question." -" Repository of information Wikipedia Spanish-English: Spanish repository in which if the translation to English exists, this will be indicated to the user.</p><p>In a broader way, the user on your mobile device generates a question of definition in mQA, the system provides mobile communication system to DQA by GPRS connection (for economy and accessibility by users in case of Mxico 4 cents per KB) , in the DQA service the questions are sent to the LabTL QA system and this searches for the answer in its collection of information (Wikipedia), the answer is validated and this is returned to the user by previously established connection to the mobile device, so you use it according to your needs This architecture pretends to demonstrate the solutions efficiency in question answering by their integration into specific scenarios by mobile devices.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Perspectives and Future work</head><p>People throughout the world are increasingly relying on cell phones and mobile devices to keep them plugged in. Obviously, search will play an ever increasing role in the evolution of mobile. When will mobile search surpass desktop search? We have been expecting better search capabilities from mobile devices for some time, and know that Asia is far ahead of North America in this respect at the current time. Today, experts discuss their views about the evolution of search in North America. And, what we are sure, is that we must continue working on this line. For this purpose, the next phase of development is the implementation of the Mobile Question Answering System for spanish and English. Furthermore, we seek the application of such search in some opportunity niches such as education.</p><p>To sum up the results expected from our architecture presented in this article are:</p><p>-Architecture presented here, unlike other proposals based on short text messages <ref type="bibr">[2]</ref> is cheaper, such as was presented in section 4. -Our proposal gives a better performance because the communication via WAP is much more reliable than that based on SMS. This is mainly due to SMS-based systems have a 80 percent certainty. While the WAP protocol provides a 100 percent reliability. -Our proposal makes use of only a servlet on the server side and a simple midlet on the side of mobile device. -Furthermore, our proposal will benefit from the availability of Spanish WIKIPEDIA.</p><p>-Finally, our proposal is based on Java Micro Edition, thus it will be independent of Operating Systems (OS).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Conclusions</head><p>A consortium of companies are pushing for products and services to be based on open, global standards, protocols and interfaces and are not locked to proprietary technologies. The architecture framework and service enablers will be independent of Operating Systems (OS). There will be support for interoperability of applications and platforms, seamless geographic and intergenerational roaming. Mobile archutecture proposed in this paper has the advantage of being adaptable to any system and infrastructure, following the current trend that mobile technologies demand.</p><p>We believe the selection of topics covered in encyclopedias like WIKIPEDIA for a language is not universal, but reflects the salience attributed to themes in a particular culture that speaks the language. Our approach also would benefit from the availability of the Spanish WIKIPEDIA and the English WIKIPEDIA.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>The distributed computing introduces new challenges because it has new resources that show very different characteristics with heterogeneous architectures connected by distinct networks.</p><p>The increasing of Internet resources has been growth up, so the necessity to manage them has been initiated. This management has become an obligation of any distributed system that search the best functionality for future applications.</p><p>It is convenient to stand out that the negotiation of the resources consist to facilitate by the optimum way the available resources on Internet, databases, big stored systems or the files systems, among others. It is to be able to predict failures, network errors, traffic and other requirements.</p><p>The decentralization is important in applications where having global information is impossible. The decentralization gives to the networks or applications a high scalability and do not allow to have a single point of failure. The peers are autonomous and in some cases they are anonymous. The decentralization generates challenges in the security for different policies of management and dynamism.</p><p>The high dynamism of the participants requires new services. These services like the self-organization, replication, searching, among others. Those tasks give performance to the distributed applications in the context of collaboration group.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.1">Content Management System</head><p>There are three important characteristics of the contents that can be used for the proposal of the management system with Peer-to-Peer (P2P) Networks:</p><p>1. The growing of the non structured data <ref type="bibr">[1]</ref>. 2. The management of the content of the network edge is considered an extension of the management of documents. The management of contents must be oriented to the reuse of the content.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">The collaboration for Internet between the users of universities is frequent</head><p>and useful in the actuality. The collaborative applications like BSCW <ref type="bibr">[2]</ref>, MOODLE <ref type="bibr">[3]</ref> like others, are client-server in the WWW, the contents is in the server and the participants depend on the robustness of the server for having the active contents. The idea of using P2P networks is not new in the management of contents. Several proposals exist for the management of the content in the companies <ref type="bibr">[4]</ref> and some applications where the content is distributed by the audio, i.e. Gnutella <ref type="bibr">[5]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">The proposal requirements</head><p>The proposal is derived to offer solutions in the scenario that is in these three suggested dimensions, Fig <ref type="figure" target="#fig_44">1:</ref> 1. The roles indicate the activity degree that the participants that have and their types of messages specifying protocols based on messages that allows guaranteeing a successful sending and reception of messages. 2. Distributes Hash Table (DHT): they are a weakly-connected and stronglyconnected this last, gives consistency the table routing. Today, the stronglyconnected DHT have the lookup with O (log N) to O (log (1)). 3. During the earlier works, the data storage had been oriented from adjust of the system index to the file replication. The cache is emphasized to save temporally, during the restoration mechanisms of the routing tables until the defined roles like in JXTA <ref type="bibr">[10]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Fig. 1. The application dimensions 3 The Simulator proposal</head><p>In this simulator a brief description is made.</p><p>1. Classes and scripts with the configuration file. The design and the operation are explained. 2. The different events that the class DHT and semantic has.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Scripts and Classes with the file of configuration explaining the design and the running</head><p>In the programming guide, the design and style of the source code of bamboo <ref type="bibr">[6]</ref> can be observed. Two stages or classes were created. The first class gives the number of hops through the DHT routing or the semantic routing. The second class is for locating the puts and generates the gets. The puts will place the objects (documents) through the bamboo nodes and the gets will find the request. Moreover, two scripts that have two classes.</p><p>The two Scripts are made in Perl. The first script is the responsible for creating the nodes type bamboo network, moreover of sending the java threads to the nodes of a cluster, balancing on each node in the cluster, the number of bamboo network nodes. For example, if there are 500 Bamboo-DHT nodes and we have ten nodes of a cluster In the article <ref type="bibr">[9]</ref>, the chum rate is made with ranks of 8/seconds to 4/minutes. This means that following the experiment, the paper shows that every 8 seconds 8 nodes are disconnected and 8 nodes are connected maintaining the session time average in 0.72 minutes with 500 nodes. The highest rank is about 4 nodes that are disconnected every minute and four nodes that are connected during the same minute maintaining a session time of 86.64 minutes.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Example:</head><p>t med = N ln 2/ where N = number of nodes, with churn-rate from 8/seconds to 4/minutes with one t med =1.4 minutes to 3 hours. In the experiment of <ref type="bibr">[9]</ref>, they have N = 1000 nodes. The estimation ln(2)/ 0.008 = 86.6433976 = 1.44 minutes. In our case N = 500 nodes, then 8/500 =0.016, ln 2/0.016 = 43.3216 = 0.7220 minutes.</p><p>In the formula ln(2)/ . Where, = (4/60) / (1000) =0.00006666666. The total is = 10397.2181 = 173.28 minutes almost three hours. With N = 500, (4/60) / 500 = 0.00013333. ln(2)/0.0001333 = 5198.603867 = 86.643397 minutes.</p><p>The second script calls the class puts-gets that is located in the class DHT that is of bamboo. The puts follow the law of Zipf and they are made from parallel form in each node. The requests are made by each node that asks for objects with a balanced form. In other words, if there are 1000 requests and if there are 500 nodes, 20 requests are sent per node and each node will follow the Zipf law with random objects that the bamboo network has. Remember than the objects were put before by the puts class.</p><p>Using the DHT class calculates the number of hops using the Bamboo-DHT routing. Using the Semantic class, the number of hops is calculated using the semantic routing.</p><p>The limits can be changed, for example:</p><p>The number of networks of the contents. It is relation with the distinct gateways of the beginning that will provide configuration file; this will generate different bamboo-network of content. This result that each bamboo network of content can make the next variations 1. The number of nodes that learn: It is relation with the total number of bamboo nodes, we choose the random number and it will tell us the nodes which will learn, these change can be done with the class DHT or semantic. 2. The number of objects: they are all the total documents that we can random put in the entire bamboo network; there the class puts/gets is found. 3. Storage in each node: During the time that the requests are made, there exist a maximum number of objects that can store each node and it's found in the DHT semantic class. 4. The requests following the Zipf law: The node that makes the request following the Zipf law. Making sure that the popularity of the objects is found in the puts/gets class.</p><p>Using the DHT class, calculates the number of hops using the Bamboo-DHT routing. Using the semantic class, the numbers of hops are calculated using the semantic routing. The example of the name_file.cfg, it is a structure in the following way: begins with &lt; sandstorm &gt; and concludes with &lt;/sandstorm&gt;. In the global parameters the node_id contains the IP of the participant; it in this case, puts the IP of the node of the cluster and the port that distinguish each node. Stages that are bamboo classes are contained. Each stage contains parameters that allow controlling certain characteristics of a physical network. Certain stages were used to allow the creation of a bamboo network like in the reference <ref type="bibr">[7]</ref>. The stage Network generates parameters of a network and the stage Rpc allows the remote procedure call. Router allows using the routes based on the bamboo-DHT considering two groups of neighbors (leaf-Set and table Routing). DataManager manages the number of replicas that can store when we use a Put. StorageManager allows keeping the data physically in the directory that is indicated. DHT calls to StorageManager. Gateway indicates that port used each node for the communication and the step of messages. The stage WebInterface allows to make call with the protocol http in a navigator to show the bamboo nodes. We conclude with StageSemantic_dht, this class allows using our algorithm which adds additional caching to each participant and allows choosing other possible routing.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Example of a configuration files in</head><p>For example: for a node that initializes the cluster that has IP, 196.0.0.1 the first node, the second 196.0.0.2 and so forth until the node 30 of the cluster, 196.0.0.30. If begins with the port number 3630. The first node 196.0.0.1:3630. The port 3631 reserves for the step of messages (gateway_port of the stage gateway of the file name_file.cfg) and the port 3632 is used to visualize in a graphic window the bamboo nodes. In ten bamboo nodes in a cluster node, would be made increasing of three in three each port and the address IP. Example: 196.0.0.1:3633, 196.0.0.1:3636... 196.0.0.1:3657.</p><p>The entire DHT networks have an origin. Bamboo uses the gateway of the stage Router, i.e. all the nodes belong to the bamboo network that is formed with the same gateway. In this case, has the port beginning of the first node 3630. If the bamboo network begins in the nodo1 of the cluster, all the nodes of our application have the following node gateway: 196.0.0.1:3630.</p><p>In the stage StorageManager to have a directory to locate the components of an object physically: value and content. Moreover, the directory /tmp is used.</p><p>This simulator allows that can be used the class java to scenarios with real nodes. It is transparent to the user and can expand the work to nodes PlanetLab <ref type="bibr">[8]</ref>. Now, the open-DHT and bamboo are in the nodes PlanetLab.</p><p>The simulator works as follows: Once that execute the first script create nodes Bamboo-DHT network. The second script put to the objects in the all bamboo network. The objects are distributed in a uniform way in the bamboo network. The following step is send the class Gets and to do the requests of the objects. For example: suppose that are 100 objects and 50 nodes bamboo, a node requests an object for the first time, the algorithm calculates the number of hops and sends a message to make the copy to the node in its caching and this guarantees that if any node sends a requests the object and if during the journey have the object then has it reduces the number of hops. If makes the request again, now reduce in smaller number of hops because now already has the object.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>3.2</head><p>The events that to have the class DHT or semantic</p><p>The events that use the DHT Routing (DR) or semantic Routing (SR) are: Log of the DHT, is when the bamboo-DHT is used to reach the object. Log of average hops of an object, the maximum of hops of an object, minimum of hops of an object, and number of times of the object.</p><p>Log that counts number of total objects for each object. In this log, helps identify how many times they are requested each object, remember, that they are objects that are located in the bamboo network.</p><p>Log that counts the numbers of total hops.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Evaluation</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Platform</head><p>The platform to measure our architecture this made in Bamboo-DHT <ref type="bibr">[9]</ref> and the language scripting is Perl; the code is within a cluster of 16 nodes. The nodes of cluster have dual microprocessors AMD Opteron 64, 1.4 GHz, connected by Ethernet Gigabit, RAM memory of 2 GB a hard disk SCSI of 36 GB and OS fedora Linux Core 6x86-64. The simulations are execution-driven. All the file logs are stored in each participant of the network overlay P2P in run time.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Evaluation</head><p>For the evaluation three modules or classes in Bamboo were added. The first module, is caching, that assigns cache in each participant-DHT and allows the node to choose between two routings. The second module makes simultaneous requests in parallel. This allows making requests of queries from different participant nodes at the same time. The third module is the replication of content. This module will consist basically of the extension of the epidemic algorithm of Bamboo-DHT to update the contents in caching of each node.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">Static Scenes</head><p>The evaluations are made both routing SR and DR. In Table <ref type="table" target="#tab_8">1</ref>, shows the parameter that was used for static scenes with Bamboo-DHT routing. The number of tests that were made was approximately 100 by each scene, although from test 45 the graphical one shows a uniform behavior with respect to the growth of the average of the number of hops. The Table <ref type="table" target="#tab_10">2</ref> shows the parameter that was used for static scenes with SR. A metadata in query of DHT is used; in addition caching that was made vary in the size of each node or participant. The number of objects is important, if are a small number of requests, 5, 10 or 20 requests. The objects will vary of 100, 250, 500, 750, and 1000 with duration of the object of one week (7*24*3600) or a day (24*3600) guaranteeing that the objects always are in the bamboo network. The communications between nodes of the cluster are stable in approximately 85% of the nodes. This guarantees 85 % of objects. In the number of requests: 10000, 12500, 15000, 20000, observed that the limits of the average are reduced. The size of caching used are: 5, 10, 15, 20. The use of caching represents use of DHT in SR. In scenes 1, 2, 3 of table 2, using small caching, would use of Bamboo-DHT. In experiments 4, 5, 6 used greater caching, would use less the Bamboo-DHT that is reflected in a smaller number of hops.  In Fig. <ref type="figure" target="#fig_1">2</ref>, the minimum, 3er quartile and the average of the number of hops by searching each object were found. The graph of Fig. <ref type="figure" target="#fig_1">2</ref> shows the following: The five more popular average objects, the total of objects represent 95% of the total requests. The requests are sent from 500 nodes of a total of 500 available nodes. Each node sends n requests of m available object. Each experiment of 500 nodes uses 500*n requests. In x-axis, the objects are ordered from the most popular to the least popular, indicated by the named percentage and with their labels from left to right. In the yaxis, the number of total hops is shown. In Fig. <ref type="figure" target="#fig_1">2</ref>, there are five most popular objects with three bars, respectively. The two first bars of each object represent semantic routing proposed and it is a combination of both routings. The last bar of each object represents Bamboo-DHT routing without alterations. The bars are divided in quartiles; the maximum number of hops for a same object is the high part of the bar. The third quartile represents the superior part of the line that is within the bar, the average or second quartile represents the high part of the black picture and the first quartile is the minimum hop that was obtained by requests of a same object that represents the final part of the line. The average of hops diminishes when the object is more popular. We observed in the graph that improvement in the number of hops by DR and SR are the two initial bars. DR represents the third bar of each popular object; the average is over the two bars.</p><formula xml:id="formula_10">------------- 2 500 250 12500 ------------- 3 500 500 15000 ------------- 4 500 750 15000 ------------- 5 500 1000 20000 -------------</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Conclusions and Future Work</head><p>The main contribution of this work is to improve the routing of the Bamboo-DHT, with semantic routing that reduces the number of hops average. The culmination of this work will have two contributions, first: 1. a mechanism of content delivery that allows possible routings to reach the content, when adding semantic routing in Bamboo-DHT, obtained a reduction in the number of hops averages to find an object and 2. A search mechanism using semantic information, giving expressivity to DHT, single-key was used and it does not affect the costs of the communication between nodes of DHT.</p><p>In future work, the last contribution will end with the class of bamboo that extends the broadcast algorithm, in addition, the evaluation in dynamic scenes, with parameters likes churn-rate and limitation of resources.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Nowadays education is leaving behind traditional teaching methods and adopting new learning techniques thus, new concepts arise as e-Learning, virtual learning environments, among others. E-Learning is the educational process placed through mechanisms supported by Internet technologies using audio, video, text, and graphics transmissions <ref type="bibr">[1]</ref>.</p><p>There are places around the world where learning techniques based on virtual worlds are highly used. Teachers and students coexisting in the same virtual environment (like a classroom), but all of them in different locations is a very interesting idea, so many projects are being developed to implement this teaching technique in near future.</p><p>There are several approaches trying to combine virtual worlds environments and elearning systems, an example of this is Sloodle <ref type="bibr">[2]</ref>, which is a fusion of Secondlife environment and Moodle platform.</p><p>Moreover, Retrieval Information area has taken great boom among researchers by the need to develop new software techniques for helping users in search processes. Usually a user loses a lot of time finding interesting information looking for quantity or quality of it.</p><p>A solution for this problem is using recommender agents. The function of these agents is to search information related to a user and, as the name involves, recommend to him.</p><p>Recommender Systems try to be one step forward from traditional information retrieve, which works by using keywords to find information about a topic through the well-known search engines (google, lycos, yahoo, etc). As the name indicates the recommender systems recommend or suggest to the users the items or products based on their preferences, they are used by Web sites, like electronic commerce, as marketing tools in order to increase the sells presenting to the user those products that user wishes (or may wish) to buy. Thus, we can know what customers like or dislike <ref type="bibr">[3]</ref>.</p><p>Nowadays many websites offer recommendations to their users, these sites are mainly focused to online sells, and of course, the recommendation systems are focused to that area too. An example of this is the E-cousal system, which is based on catalogue sells and supported by a web page and an application functioning as a multi-agent system. <ref type="bibr">[4]</ref>.</p><p>In virtual worlds, users are represented by avatars and they can interact among each other in a 3D world. In those environments there are different services offered to them, with a drawback: the larger a virtual world becomes, the harder it is to find new services or activities added to it (just like the real world: the larger a city becomes, the harder it is to find certain places), these services could increase in a chaotic and uncontrolled way, causing serious limitations for their management, organization and retrieving.</p><p>The paper's proposal is to show the advances in the developing of an agent in a virtual world; showing how is built the user profile and the implementation of different modules of the system.</p><p>In Section 2 we describe architecture of the agents; in Section 3 we talk about the services and how they are managed; in Section 4 we show the implementation of 4 modules: database, the user profile, the algorithm for rating recommended services and some learning rules; in section 5 we describe some results and finally in section 6 the conclusions and future work.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Architecture</head><p>The main components of the recommender agent are shown in Figure <ref type="figure" target="#fig_0">1</ref>  <ref type="bibr">[5]</ref>. Components are listed on rectangles whereas the exchange of information is listed with arrows unidirectional or bidirectional depending on the function. All of these run under OpenSim platform, which is an open source virtual world simulator <ref type="bibr">[6]</ref>. In this work we explain in detail how we build the user profile, the Database implementation, and part of the learning techniques, an improvement of the algorithm for rating services and how the services used by the recommender agent are managed. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Services Module</head><p>The services in a virtual world could be varied, from recreational activities to debates or discussions about a specific topic. Services are divided into these categories: Introductory courses, Conferences and Interactive talks, Classes, School courses, Recreational courses, and Parties, luncheons and events.</p><p>At this moment the creator of a service is responsible to introduce the service information in the database, filling the required information in a Web page. The data required are: service description, title, keywords and type of service among others. The recommender agent needs this information in order to recommend services. The options for the service creator are shown in Figure <ref type="figure" target="#fig_1">2</ref>.  The user access to their services through a web page bases on Apache server, it allows to create, modify, or delete a service related to that user. When a user creates a service the data required is inserted as a new row in the database and it's related to the user, if the user modifies a service, previously created, the information is updated on the database. The user also can delete a service, in this case the row is not deleted from the database, and the service is only marked as disable.</p><p>Using this page, users can offer their services and they can be recommended to other users. Also this page allows the management of services related to a user.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Implementation</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Database Implementation</head><p>OpenSim simulator, mentioned above, includes a database which is essential for simulator's proper working, and in which users, objects, regions and scripts' registration take place in the virtual world. The database has a table named users, in which user IDs (UUIDs) are defined, those IDs are used to link original database tables with new tables needed by the recommender agent in order to meet their goals.</p><p>Figure <ref type="figure" target="#fig_2">3</ref> shows how the tables are linked to the existent users table, in one hand it's linked to the user profile table, which in turn is connected to the topics that the user prefers. In the same way, the activities are related to a user by his ID. Finally we store the recommended services' list for subsequent rating. able Profile stores the user information, it has a foreign key UUID (users), that is be Topics store the keywords associated to services, these keywords are the the Prefers connect the table profile and the table topics, it has a primary key wh es the services created by users and registered in the services we tween the user and an object (which can be users, only ha mendations and Services, it connects the se eives, which connects a recommendation list to a user Its pr</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Building user profile</head><p>Sensors are objects located along the virtual world, by which they listen the ctivities in the database, ensuring availability fo e show the interaction of sensors, the preferences classifier agent and the T cause profile is just an extension of the user information, but the table users is already defined in the original Opensim database and we didn't want to modify those tables.</p><p>Table <ref type="table">mes</ref> that user prefers. This table has its primary key TID, the name of the topic and its rate.</p><p>Table ich references to UUID (profile) and TID (topics). This way we can now what topics a user prefers.</p><p>Table <ref type="table">Services</ref> stor b page, its primary key is SID, and has more information as created, description, region, in order to know where the object is.</p><p>Table Activities registers any interaction be linked to a service), it has a primary key AID, and two foreign keys, UUID (users) and SID (services). These activities are not stored permanently in the database, once the activity has been analyzed, it is marked as "checked" and then deleted.</p><p>Table Recommendation stores the recommended services list sent to the s its RID and the date the list was sent. communication channel used in OpenSim (registers interaction between avatars and objects associated to services in simulator). These sensors are programmed using Linden Script Language and the information retrieved by them is sent to a web page in order to store that data in the database, this is because the language is very simple and it doesn't allow connection to databases or output of information into a file.</p><p>Sending the data to a web page is very helpful, and using PHP (which is a robust language) we can connect to databases. Sensors register information in table a r future analysis. This table is continuously handled by the preferences classifier agent to debug information stored in it. The agent reads this table, and according to the information read, inserts or increases the rating in table topics. After doing this, the agent delete the information previously read in order to avoid garbage data stored in the database.</p><p>In figure <ref type="figure" target="#fig_3">4</ref>, w tables in the database. This is part of the architecture, but with more details showing the process to build the user profile. Finally it is worth noting that the sensors inside the virtual world are in a passive fo</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">Algorithm for rating recommended services</head><p>Tasks related to the preferences classifier agent are divided in two phases: Phase 1 atabase, where services, which avatar ha se new values for the ratings of recommended services in the Algorithm 3.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>us lgorithm 3 ent retrieves services rated by the user in table recommendations table</head><p>If rating associated to service is: rm, they only changes to an active form when an avatar is close to them, and then, they starts to collect the information.</p><p>Activities analysis; Phase 2 Rating analysis <ref type="bibr">[5]</ref>.</p><p>In phase 1, the agent continuously monitors d s interacted with, are stored, it classifies the services by preference (p) and by length (l). Preference is measured by services accessed frequently by the avatar, and length is measured by the elapsed time since the avatar started to use the service until it finishes using it. The agent applies Algorithm 2 in this phase and Algorithm 3 on phase 2 <ref type="bibr">[5]</ref>.</p><p>We propo With previous values the rating just increase its value showing only positive preferences, nevertheless using that values we can't know if the user dislike that service or it just doesn't matter to him. Moreover a problem arises just by increasing the rating: "is there a bound for the rating value or could it increase without a limit?"</p><p>The previous values were: increase 0.1 if the user rating is 1, increase 0.2 if the er rating is 2, same for rating on 3, 4 and 5, if user rates 0 do nothing. Below we show the algorithm with the new values:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A</head><p>Step 1: The ag Step 2: Select keyword from table services in order to update or insert it in topics Step 3: 0: Decreases rating in 0.3</p><p>These rules are very simple but solve the posterity problem explained above. More ru</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Results</head><p>Web page access to management of services is working properly and there has not nication channel and we are doing so n and the learning rules have not</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Conclusions and Future Work</head><p>This paper shows an important advance in the implementation of the recommender op References 1. Marcus, A. "Caracterización de los Servicios de la Educación a Distancia desde la óptica de 2.</p><p>ne, Judith Kemp, and Paul Andrews. Sloodle Learning System for Virtual 3. ue desde los les will be developed for working analyzing user activities and updating the user profile had problems. The page is based on Apache server using PHP and MySQL. Services related to users has been added and stored in the data base, and the modification or deletion of them works fine too. Relations of the tables in the database fulfill its role, and no problems have been reported related to it.</p><p>Sensors register information read on the commu me tests to ensure that the web page using PHP stores the data efficiently in the database, one disadvantage of using web page is that the server may be saturated with connections, but at the moment this is the best solution.</p><p>The Learning Techniques module is under constructio been implemented yet, but we hope shortly have some results to show. agent, step by step the different modules are been developed, and they works properly. There has had some complications developing the sensors but they has been successfully solved bypassing the language and the database through a PHP webpage.</p><p>Currently the user profile is being developed and analyzing it in order to find the timal balance between the user preferences and user activities. Agent learning capacities are based on deductive rules, which are being created. Only the rules presented in this work have been developed, but when all the rules are ready, the agent could infer the best services to recommend to the user. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>This paper has been constructed on the basis of vast experience gained through several years on the characterization of the academic language (record/register). Moreover, this research is the result of a deeper study on postgraduate careers of a public university (BUAP), which let us to detect the most frequent mistakes and transgressions on the academic language. We will start this discussion with two questions: 1) What is the meaning of those terms that we hear and use, but whose conceptualization or definition will state difficulties of some kind? In academia, especially in recent years, there are some terms that are frequently used within institutional development plans. Terms such as: pedagogy, didactics, methodology, teaching techniques, teaching process learning, etc are used as a prelude of academic excellence and resource optimization (human resources and materials). And here becomes the second question: Does not imply knowledge any kid of specific knowledge, a specific language, a certain jargon? If we accept this assertion as a claim we may infer that by acquiring a scientific knowledge 1 of language (knowledge instrument 2 ) we can facilitate the language knowledge and application of the same. But the real question is the following: Does the use of a semantic Web require some user verbal skills or at least to advise the lack of these skills?</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Considerations</head><p>Errors or deficiencies affect at different levels the following issues: intelligibility (Rules 1, 2, 3 of verbal system); the thought logic (Rules 3, 4, 5); communication (null or minimal noise in the process of encodingdecoding); speech fluidity and precision (speech alludes to both phrasing and reasoning). Natural language is a system of verbal symbols (sorted set of functioning rules of constituent items). The academic subsystem (record/register) and jargon (characteristic of the different areas of knowledge) is characterized by the following rules.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Phoneticphonological and spelling rules</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>•</head><p>Signs Graphics: punctuation and auxiliary signs.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>•</head><p>Spelling: diacritics, punctuation, accentuation, orthosemantic problems.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Lexicalsemantics rules</head><p>Lexical (3rd unit of language, 1st significant unit): univocity, precision, accuracy, conciseness, versatility and richness vs. ambiguity, categorization error, polysemy, synonymy, homonymy, homophony, paronymie, imprecision, vagueness, semantic phrases, poverty, repetition, ignorance, confusion.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3">Morphosyntactic rules</head><p>• Constructive (5th unit of language, 1st logical unit): logic, consistency, sequential, cohesion, consistency, clarity, simplicity vs. illogical (juxtaposed and copulative; alteration of conjunctive nexus, etc.) assumptions, fractures, anacolutha, truncated sentences, inconsistency (especially subjectpredicate), confusion (especially in terms of managing the elements of the compound sentence) , scavenging, and Pseudocultism (ex, abuse of "el cual" which); incorrect temporal correlations, and so on.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>•</head><p>Concordance, among others.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.4">Encoding rules</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>•</head><p>Encoding: denotative, which corresponds to the formal record/register (academic, scientific, technical) vs. subsystem of use, characterized by a tendency of connotative encoding, the standard or colloquial register, oral, with semantic shifts or meanings tagged by chrono and geo sociolects -the use of archaisms (a word or form of speech no longer in common use); or only functional in some areas or among certain social groups -altered prepositional regimen, etc. We are particularly interested on those problems related with the word unit. Therefore, we will not stress on unacceptable underlying assumptions, stemmed phrases, anacolutha and diverse fractures, erroneous punctuation, with the resulting disruption of the correct decoding or in cases of disagreement, but we will focus our attention on aspects related with Rules 1 and 2: Phoneticphonological and spelling rules; Lexicalsemantics rules: transgressions and most frequent errors and we will present some examples.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.5">Logicalstylistic rules</head><p>• Vocabulary deficiencies and use meanings 3  • Confusion/alteration of logicalgrammatical category • Spelling errors that affect meaning (written accent mark: íntegrointegrointegró; gráficagrafica, etc.). • Problems with homophones, homonyms, paronymie and polysemy <ref type="foot" target="#foot_12">4</ref>• Abuse of some Spanish verbs when defining and writing, such as: ser(being), estar(to be), tener(to have), poder(to be able), hacer(to do). • False friends from the English language (Due to the users read literature in English: remover (remove) instead of eliminar, interface or interfase (interface) instead of interfaz, etc.). • Abuse of lexical and constructive Anglicism and the problems they cause 3 In order to simplify and due to we are interested on the determination of the dialectic phenomena, we create the concept of use subsystem. We will focus on transgressions with respect to the academia norms that rule this record/register.</p><p>• Misuse of the gerund (English gerund) • Forms of 'passive English' (construction of the auxiliary 'to be' (ser) rather than reflected passive)</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Experimental results</head><p>In this section we present an evaluation of a written test carried out on 135 students of different postgraduate programs on science and technology. We asked them to define different words. Even if the complete list was conformed by 25 terms, we only show the obtained results with a subset of this set. In summary, we have detected 13 problems, which are described as follows.</p><p>1. Ignorance of the term meaning (null answer). Surprisingly, not one word were answered in blank (null answer), even on terms such as vaso(glass), sección(section), adolescente(teenager), where answer percentage rates were 1,48%, 2,22% and 9,62%, respectively. Percentages rates reach values such as 67,4% (acerbo), 65,92% (asechar) and 50,37% (sito).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>•</head><p>The pair acerboacervo, had 91 and 51 null answers (67,4% and 37,7% respectively), from which only 2 and 46 where partially correct (2,2% and 34%).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>•</head><p>The pair adolecienteadolescente, had 34 and 13 null answers (25,1% and 9,6%), from which 37 and 69 were correct (from this 37, only 5 were totally correct, i.e., those that gave more than one meaning; in the second case, i.e., adolescente, there is only one meaning which may be expressed in different manners).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>•</head><p>The pair basobazo, had 66 and 34 null answers (44,8% and 25,1%), from which 29 and 93 where correct (partially correct in general, since in major cases there were given only one meaning) with percentages of 21,4% and 68,8%, respectively. The above presented examples suggest taking some considerations in order to avoid that the requested information cannot be retrieved due to this type of user language fails. We do not expect an information retrieval system tries to "guess" what the user wanted to say from his written query (what he really said).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Confusion</head><p>In the digital version of the Spanish Language Dictionary (Diccionario de la Lengua Española, RAE) there exist a simple mechanism of word 'approximation', i.e., if someone writes a word incorrectly (ex. 'grafica'), it suggest the correct word ('gráfica' or 'graficar'), i.e., it looks for orthographical similar words, which we consider to be a good example of semantic Web.</p><p>As conclusions, let us consider the following issues: a) The language, any type of language, is a symbol system. A system is an ordered set of functional rules of the system items. In this case, we mean as symbols as those that refer to something, which in general is different of itself <ref type="foot" target="#foot_13">5</ref> and with intentionality <ref type="foot" target="#foot_14">6</ref> . b) The classical analysis <ref type="foot" target="#foot_15">7</ref> of the verbal sign demonstrate the use of three elements:</p><p>the phoneticacoustic or significant (and its graphical equivalence, i.e., the letters); the eideticconceptual or meaning, and the Referent or reality, i.e., the target object or what it is referred (it means, it denotes). c) The concept, the idea constitutes the basis of the logic and abstract thought; together with the judgment and argumentation, the complex. The judgment brings together some concepts under certain predication rules (the sentence is the subject predication), and the argumentation, judgments.</p><p>d) The thought is defined as rational speech, and speech is the language as act. It is an update of the potential or human ability of thinking and talking, which implies to learn the reality and not only to extern it (communicative function). e) The knowledge is the apprehension of Reality, and this, which includes the principle of identity and spacetemporal coordinates, may be defined as the set of all referents. f) Such apprehension of Reality is performed through the language. It relies under the eideticconceptual element, in abstraction and generalization of referents (concrete or abstract, folkloric, geometric, cultural or literary), its reduction to the essential characteristics (different to the accidental ones). g) Finally, but no by coincidence, if we follow the Aristotelian logic and categorical grammar, we will find the overlap between grammatical and logical categories, particularly between ser(to be)sustantivo(noun); accidente(accident) adjetivo(adjective); acto(act)verbo(verb).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Conclusions</head><p>On the basis of the above considerations presented (the thoughtlanguage process). This process will depend on the level of system dominion, i.e., to think and link related concepts implies a lexical database and the management of rules in order to link them (grammar). We do not think with images or icons, but with concepts, ideas. For a simple test, we suggest the reader to think (not to imagine) in a certain reality (abstract or concrete, physical or mental) without appeal to the language (language of any kind) that is no using words. This paper presents the results of an analysis of the representative works related to the reseach lines of k-means algorithm devoted to overcome its shortcomings. To establish a framework for a proposed improvement to the standard k-means algorithm, the results obtanained from experiments of the kmeans in the Matlab package and databases of the UCI repository are presented.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Research issues on K-means</head><p>Keywords: k-means, clustering, k-means-Matlab.</p><p>1 Introduction.</p><p>The problem of object clustering according to its attributes has been widely studied due to its application in areas such as machine learning <ref type="bibr">[4]</ref>, data mining and knowledge discovery <ref type="bibr">[3,</ref><ref type="bibr">11]</ref>, pattern recognition and pattern classification <ref type="bibr">[2]</ref>. The aim of clustering is to partition a set of objects which have associated multidimensional attribute vectors into homogeneous groups such that the patterns within each group are similar. Several unsupervised learning algoritms have been proposed which partition the set of objects into a given number of groups according to an optimization criterion. One of the most popular and widely studied clustering methods is K-means <ref type="bibr">[10]</ref>.</p><p>This work is part of a project devoted to proposals for improvement to k-means algorithm and its application to health care in México. In the works of <ref type="bibr">[29]</ref> and <ref type="bibr">[30]</ref> proposes an improvement to k-means algorithm using a new convergence condition and application in cancer incidence by municipalities in México is proposed.</p><p>Currently, the project focuses on making a proposal for improvement of the algorithm based on the optimization of the classification step. In doing so, we have done an update and analysis of the state of the art in theoretical research of the algorithm. For purposes of establishing a framework to propose an improvement, a serie of experimental tests on the matlab package for kmeans in databases of the UCI repository has been made.</p><p>The works of <ref type="bibr" target="#b127">[41]</ref> and <ref type="bibr" target="#b128">[42]</ref> highlight the impact and relevance of the k-means algorithm, within the context of Data Mining and clustering techniques. There are over a thousand papers published to date, related to its different forms, applications and contributions of k-means algorithm. Research on the k-means algorithm, has been developed in two major directions, from theory and field of applications. From theory, a set of advantages and shortcomings has been identified, in an attempt of try in to overcome their shortcomings. Also, a serie of works have been developed, that have led different lines of research, on wich this work focuses. Study we have made no attempt to be exhaustive, but rather representative of the reseach. We include in this report, selected works that have been frequently cited and work that was deemed to provide a novel approach. This report is organized as follow: following this introduction, it is included in the Section 2, the approach to the clustering problem and the description of the k-means algorithm. The section 3 provides an analysis and synthesis of the works in the various research lines of kmeans. Section 4, describes certain characteristics of k-means in Matlab and in the section 5, we present a reference database of UCI repository and results of test performed to illustrate some aspects of the performance of k-means algorithm on Matlab. Finally, in section 6, we will discuss our preliminary results and present the future work.</p><p>2 Clustering Problem and the k-means Algorithm.</p><p>According to <ref type="bibr">[15]</ref>, clusters analysis aims at solving the following very general problem: given a set X of N entities, often described by measurements as points of the real d-dimensional space , find subjets of X wich are homogeneouns and/or well-separated. Homogeneity means that entities in the same cluster must be similar and separation between entities in the same cluster must be similar and separation between entities in different clusters must differ one from the other. These concepts can be made precise in a variety of ways, wich lead to as many clustering problems and even more heuristic or exact algoritms, so clustering is a vast subject.</p><p>The work of <ref type="bibr" target="#b128">[42]</ref> provides a good survey on the issue of clustering and provides a set of key references.. According to <ref type="bibr">[15]</ref>, a mathematical programming formulation of the minimum sum-ofsquares clustering problema is a follows:</p><formula xml:id="formula_11">min f(M,Z) = 2 subject to = 1, j = 1,2,…,N € i = 1,2, …, M; j =1,2, …,N where = , i = 1,2, … M.</formula><p>The N entities to be clustered are at given points = (x i1 , x j2 , …., x jd ) of R d for j = 1, …., N; M cluster centroids must be located at unknown points m i € R d for i = 1, …, M. The decision variable z ij is equal to 1 if point j is assigned to cluster i, at a squared Euclidean distance 2 from its centroid. It is well-known that condition € may be replaced by € , since in the optimal solution, each entity belongs to the cluster with the nearest centroid.</p><p>Among the clustering algorithms based on minimizing an objective function or the squared error, perhaps the most widely used and studied is called the k-means algorithm. This algorithm has been discovered by several research across different disciplines, most notably <ref type="bibr">[23]</ref>, <ref type="bibr">[24]</ref> and <ref type="bibr">[12]</ref>. And the best-known heuristic for minimum sum-of-squares clustering is MacQuee`s <ref type="bibr">[24]</ref>. It proceeds by seleccting a first set of M points as candidate centroid set, then alternately (i) assigning points of X to their closest centroid and (ii) recomputing centroids of clusters so-obtained, until stability is attained.</p><p>In a more specific form in the work of <ref type="bibr">[29]</ref> four algorithm steps can be identified:</p><p>Step 1. Initialization. A set of objects to be partitioned, the number of groups and a centroid for each group are defined.</p><p>Step 2. Classification. For each database object its distance to each of the centroids is calculated, the closest centroid is determined, and the object is incorporated to the group related to this centroid.</p><p>Step 3. Centroid calculation. For each group generated in the previous step, its centroid is recalculated.</p><p>Step 4. Convergence condition. Several convergence conditions have been used from which the most utilized are the following: stopping when reaching a given number of iterations, stopping when there is no exchange of objects among groups, or stopping when the difference among centroids at two consecutive iterations is smaller than a given threshold. If the convergence condition is not satisfied, steps two, three and four of the algorithm are repeated.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Analysis and Classification of Papers Reviewed.</head><p>This section, identifies the advantages of k-means. The following is the result of the analysis of the work that has been developed in different research lines of k-means, related to the proposals made to try to overcome their shortcomings. The results are organized by category it, has been included in each case, the authors' names, the title and a concise comment on the thrust of the work.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Algortihm K-means Advantajes.</head><p>MacQeen J. <ref type="bibr">[24]</ref>, the author of one of the initial k-means algorithm and the most frequently cited, states.</p><p>The process, which is called "k-means", appears to give partitions which are reasonably efficient in the sense of within-class variance, corroborated to some extend by mathematical analysis and practical experience. Also, the k-means procedure is easily programmed and is computationally economical, so that it is feasible to process very large samples on a digital computer.</p><p>Likewise <ref type="bibr" target="#b125">[39]</ref>, summarizes the benefits of k-means, in the introduction to his work: K-means algortihm is one of first which a data analyst will use to investigate a new data set because it is algorthmically simple, relatively robust and gives "good enough" answers over a wide variety of data sets.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Algorithm K-means Shortcomings.</head><p>Taking as a framework and as an extension and update, the k-means shortcomings that are identified in <ref type="bibr" target="#b128">[42]</ref> the following is the result of the analysis previously cited in a series of tables grouped by category of work that arose as extensions k-means or as possible solutions to one or more of the limitations that have been identified above.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.1">The algorithm's sensitivity to initial conditions:</head><p>The number of partitions, the initial centroids.</p><p>According to <ref type="bibr" target="#b128">[42]</ref> there is a universal and efficient method to identify initial patterns and the number k of clusters. In <ref type="bibr" target="#b126">[40]</ref> briefly is discussed the sensitivity of the algorithm for the allocation of initial centroids, that in practice the usual method is to test iteratively with a random allocation to find the best allocation in terms of minimizing the total squared distance. However, there have been various investigations aimed at making various proposals related to these limitations: Math Works Inc. (Natick, MA). <ref type="bibr">[1]</ref>. The K-means algorithm for clustering is in the following MATLAB function:</p><p>[IDX, C, SUMD, D] = KMEANS(X, K) This function partitions the points in the N-by-P data matrix X into K clusters. This partition minimizes the sum, over all clusters, of the within-cluster sums of point-tocluster-centroid distances. Rows of X correspond to points, columns correspond to variables. KMEANS returns an N-by-1 vector IDX containing the cluster indices of each point. By default, KMEANS uses squared Euclidean distances. The K cluster centroid is located in the K-by-P matrix C. The within-cluster sums point-to-centroid distances in the 1-by-K vector sumD. Distances from each point to every centroid in the N-by-K matrix D. It may include optional parameters to specify distance measure, the method used to choose the initial cluster centroid positions, display information.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Test Results of K-means in Matlab.</head><p>Tests for k-means in Matlab, used the well known UCI Machine Learning Repository <ref type="bibr" target="#b123">[37]</ref>. The UCI Machine Learning Repository <ref type="bibr" target="#b123">[37]</ref> is among other things, a collection of databases, which is widely used by the research community of Machine Learning, especially for the empirical algorithms analysis of this discipline. The Iris Data Set is a database of types Iris plant, which has No. of instances: 150 (50 in each class), No. of attributes: 4 (Sepal length, Sepal width, Petal Length, Petal width) No. of classes: 3 (Hedges Iris, Iris versicolor, Iris virginica). One class is linearly separable from the other two; the latter are NOT linearly separable from each other. Based on data and classes defined in <ref type="bibr" target="#b123">[37]</ref> and <ref type="bibr" target="#b128">[42]</ref>, fig. <ref type="figure" target="#fig_0">1</ref>  The "Test I4" is an example of the test results on Matlab and kmeans for the Iris data set. Iter column represents the number of iteration, the phase indicates the algorithm phase, num provides the number of exchanged points, sum, provides the total sum of distances, inter% are the percentage of exchanged points in each iteration.      With regard to exchanges between groups that the algorithm makes in the tests conducted, it was observed that the most significant changes occurred from first to second iteration. For all cases, in the first step all points are located (100%), the third column includes the number of points to be exchanged in the second iteration and the fourth column is the percentage difference in the number of items exchanged between the first and second iteration.</p><p>According to the results in </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Image Classification by Texture Segmentation using GAF-SVM</head><p>Sergio Manuel Dorantes, Manuel Martín Ortiz, María J. Somodevilla, Jesús Lavalle Martínez, Ivo H. Pineda Torres Facultad de Ciencias de la Computación, BUAP sergiomanuel@hotmail.com, {mmartin, mariasg, jlavalle, ipineda}@cs.buap.mx Abstract. Due to the amount of visual information that currently exists, there is a need to classify it properly. In this paper we present an alternative dual method for image categorization according to their texture content defined as GAF-SVM, this method is based in the use of Gabor Filters (GAF) and Support Vector Machine (SVM). To perform the image classification we rely on filtering techniques for feature extraction mixed with statistical learning techniques to perform the data separation. The experiments were carried out by taking a set of images containing coastal beach scenes and a set of images containing city scenes. A feature vector is obtained from applying a bank of Gabor Filters to the input images; the output feature space is then used as an input to the SVM Classifier. The Support Vector Machine is responsible for learning a model that is capable of separating the sets of input images. Experimental results demonstrate the effectiveness of the proposed dual method by getting the error classification rate to near 9%.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>I. INTRODUCTION</head><p>The proposal for an alternative method of image classification requires analysis of the methods presented so far concerning the area. Extracting visual information from an image to obtain their most important features is essential for classification tasks; over the years various approaches have been presented regarding this field of study such as: color histograms, region-based classification and gray-level values of raw pixels; though one solution has been to incorporate the texture analysis as main feature descriptor. This is largely due to the fact that most surfaces on images contain some kind of texture. In the recent years, texture analysis has been used for object recognition, image interpretation, image segmentation and classification <ref type="bibr">[1,</ref><ref type="bibr">2,</ref><ref type="bibr">6,</ref><ref type="bibr">8,</ref><ref type="bibr">9,</ref><ref type="bibr">10]</ref>.</p><p>In recent papers like <ref type="bibr">[4,</ref><ref type="bibr">6,</ref><ref type="bibr">7,</ref><ref type="bibr">10,</ref><ref type="bibr">11,</ref><ref type="bibr">12]</ref>, texture is been understudy in an isolated manner to evaluate the performance of the proposed algorithms, in some cases it has been used artificial textures, which limits the application area of these methods. Textures are used by the human visual system to separate different objects within scenes as well as surface analysis <ref type="bibr">[11]</ref>. Texture can be recognized as an irradiation patterns that are perceptually uniform. Textures can be explained as an efficient measure to estimate the structural differences of orientation, roughness, smoothness or regularity between different regions of an image <ref type="bibr">[14]</ref>.</p><p>But bring out a formal definition of what a texture really is, it became a subjective topic. As it was mention in <ref type="bibr">[13]</ref> the definition of texture is dependent on the purpose for which it is being used and outlines some definitions: 1. The basic pattern and repetition frequency of a texture sample could be perceptually invisible, although quantitatively present. In the deterministic formulation texture is considered as a basic local pattern that is periodically repeated over some area. 2. An image texture may be defined as a local arrangement of image irradiances projected from a surface patch of perceptually homogeneous irradiances. 3. Texture is characterized not only by the grey value at a given pixel, but also by the grey value 'pattern' in a neighborhood surrounding the pixel.</p><p>Our proposal is based on the use of natural textures in real world images, for that reason the classification model must deal with more complex images in natural conditions.</p><p>The 2-D Gabor filters (2D-GF) have certain properties that make them suitable for textural identification in many ways: 2D-GF have tunable orientation and radial frequency bandwidths, tunable center frequencies, and optimally achieve joint resolution in space and spatial frequency. The demodulated Gabor channel envelopes generally contain only low spatial frequencies which are optimally localized in both domains <ref type="bibr">[16]</ref>.</p><p>Gabor filter based methods have been successfully applied for a variety of machine vision application, such as texture segmentation <ref type="bibr">[10,</ref><ref type="bibr">11,</ref><ref type="bibr">12,</ref><ref type="bibr">15,</ref><ref type="bibr">16,</ref><ref type="bibr">18]</ref>, texture classification <ref type="bibr">[9,</ref><ref type="bibr">13,</ref><ref type="bibr">19]</ref>, iris recognition <ref type="bibr">[21,</ref><ref type="bibr">22,</ref><ref type="bibr">23]</ref>, on-road vehicle detection <ref type="bibr">[17]</ref>, fingerprint classification <ref type="bibr">[20]</ref>, and as mentioned in <ref type="bibr">[15]</ref> edge detection, object detection, image representation, and recognition of handwritten numerals.</p><p>This paper is organized as follows: in section II it is mention the related work we based on to develop this article, in section III it is made a detail description of the proposed method, in section IV it is an explanation of the way the input data is processed as well as the Gabor filter's parameters selection, in section V the details of the SVM classifier parameters, and in section VI the experimental results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>II. RELATED WORK</head><p>The classification of images has been studied from various approaches, most of all through the mixing of methods, one for texture extraction and one for the classification process.</p><p>In <ref type="bibr">[9]</ref> is emphasized the use of Gabor filters as a texture extraction method and classification is performed with maximum likelihood method for the classification of aerial and satellite digital images. In <ref type="bibr">[3]</ref> is proposed a method of image classification using as an image representation their color histogram and as method of classification the Support Vector Machine. In <ref type="bibr">[4]</ref>, is not used an external feature extractor, instead the SVM classifier receives the grey level values of each pixel on the image, trying to prove that SVM can implement feature extraction methods within its architecture, this method is computationally expensive due to the number of regions that can define an image. Another approach is performed in <ref type="bibr">[5]</ref> where a modification of the SVM is used for identification of regions among a group of images. In <ref type="bibr">[6]</ref> the SVM is combined with the Discrete Wavelet Frame Transform for the classification of images of the Brodatz album. In <ref type="bibr">[7]</ref> is mixed the use of wavelet transform as a feature extractor known as the pyramid-structured wavelet transform and SVM as the classification method. In <ref type="bibr">[8]</ref> is proposed a method called Gaussian Mixture Model mixed with Independent Component Analysis (ICA) to perform the image classification, which is called ICA Mixture Model.</p><p>The first step to complete the proposed method is to extract the texture features with a bank of Gabor filters applied to each input image, and then take the filter's output to form a training dataset to feed the SVM classifier.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>III. PROPOSED METHOD</head><p>In order to accomplish the image classification we rely on filter based techniques to perform texture feature extraction mixed with statistical learning theory techniques to achieve the image data separation. Gabor filters were selected to extract texture features from images due to their resemblance to the human visual system <ref type="bibr">[13]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A. Gabor Filters</head><p>A number of authors have used a bank of filters to extract local images features <ref type="bibr">[10,</ref><ref type="bibr">11,</ref><ref type="bibr">16,</ref><ref type="bibr">19]</ref>. Different authors used different sets of Gabor Filters, from spatial domain to frequency domain.</p><p>A 2-D Gabor filter is a linear filter whose impulse response is defined by a harmonic function multiplied by a Gaussian function. In the spatial domain can be defined as follows: </p><p>Where f is the central frequency of the filter,θ the rotation angle of the Gaussian major axis and the plane wave, γ the sharpness along the major axis, and η the sharpness along the minor axis (perpendicular to the wave). In the given form, the aspect ratio of the Gaussian is λ=η/γ. This four parameters (f,θ,γ,η) define the shape of the filter, and by changing them we can detect different textures. The normalized 2-D Gabor filter function has an analytical form in the frequency domain.</p><formula xml:id="formula_13">′ ′ − − + ′ = + ′ = − + Ψ =<label>(2)</label></formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>B. Filter Design vs. Filter Bank</head><p>There exist two aspects regarding the implementation of Gabor filters, on one hand the filter bank approach and in the other hand the filter design approach <ref type="bibr">[25]</ref>. In the first one, a bank of filters is formed by grouping multiple filters tuned at different frequencies and different orientations. The decision of the parameters setting depends on the type of texture to be analyzed. The difficulty of using the filter bank approach relies on the fact that their parameters are established ad hoc and are not optimal for a specific processing task. One of the goals of this work consists on presenting results that would help to specify such parameters. Furthermore, if the bank handles many frequencies and orientations, resulting in a large bank with a lot of filters within, this translates in a large number of convolutions. The filter design approach focuses on designing one or a few filters for a particular application in an effort to reduce the difficulty provided by the filters bank and also to reduce the dimensionality of the output, as well as the processing cost. The disadvantage of this approach lies in the limitation of the tasks for which it was designed. When working with a single filter it is possible that some of the textures in the images are not identified or detected as the filter has a narrow range capacity to detect local texture features.</p><p>A filters bank allows the analysis of an image in a single pass way at several frequencies and in several orientations at once. According to the characteristics of our model, the use of a filters bank is the solution choice of deployment, although it could mean an increase, in the computational processing, this is not significant. The design of a Gabor filter bank consists, in general, in the selection, for each filter, of the proper values of the following parameters: frequency, orientation, γ andη, the last two parameters known as the smoothing parameters <ref type="bibr">[26]</ref>.</p><p>In this research it is defined a bank with up to 3 orientations and up to 2 frequencies, resulting in a bank with maximum 6 output filters, allowing us to accurately detect a texture among a large set of images. This decision was made based on the studies presented in <ref type="bibr">[26]</ref>, where is compared with various parameter selection approaches, and summarizes some parameter values adopted in literature.</p><p>Using many different orientations and scales (frequencies) ensures invariance; objects and some textures can be recognized al various different orientations, scales and translations <ref type="bibr">[27]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>C. Support Vector Machines</head><p>Support Vector Machines (SVM) were introduced by Vapnik as a powerful learning tool based on statistical learning theory, a Support Vector Machine is a binary classifier that makes its decision by constructing a linear decision boundary or hyperplane that optimally separates data points of the two classes in feature hyper space and also makes the margin maximized <ref type="bibr">[20]</ref>.</p><p>SVM starts from the goal of separating the data with a hyperplane, and extend this to non-linear decision boundaries using the kernel trick <ref type="bibr">[29]</ref>. A hyperplane can be defined as:</p><formula xml:id="formula_14">0 T w x b + =<label>(3)</label></formula><p>Where x represents a point (a vector), w represent the weight (also a vector). We want to choose w and b to maximize the margin, or distance between the parallel hyperplanes that are as far as possible while still separating the data. The hyperplane must separate data such as: 0 for all of a class 0 for all of another class </p><p>If data are separable in this way, there will probably be more than one way to do it. Among all the possible existing hyperplanes, SVM selects the one in which the distance between the hyperplane and the closest data is the widest possible <ref type="bibr">[29]</ref>.</p><p>When working with a dataset that is not linearly separable, it is necessary to turn to the use of a kernel function. The kernel function allows the SVM to form non-linear boundaries <ref type="bibr">[29]</ref>. Data representation through kernel function offers an alternative solution to the nonlinearity problem, projecting the information to a higher dimension feature space <ref type="bibr">[28]</ref>. This is accomplished by changing the representation of the function; this is similar to mapping the input space X to a new space H, called feature space, in the form: :</p><p>R d X H φ ⊂ →</p><p>(5)</p><p>Now instead of considering the input vectors {x 1 ,…, x n } it is considered the transformed vectors {φ(x 1 ),…, φ(x n )} as shown in figure <ref type="figure" target="#fig_0">1</ref>. By doing this substitution, it is obtained a SVM raised in a new space (this is called the 'kernel trick'), it is important to mention that in practice the implementation of this nonlinear technique consumes the same amount of computational resources of its linear equivalent. The general problem that SVM want to resolve is to search, for a given learning task, with a finite amount of data, an appropriate function that helps to carry out a good generalization, which results from a proper relationship between the accuracy achieved with a particular training set and the ability of the model <ref type="bibr">[30]</ref>.</p><p>The use of the 'Radial Basis Function' (RBF) kernel is based on the fact that this kernel is basically suited best to deal with data that have a class-conditional probability distribution function approaching the Gaussian distribution, like the texture present on the input images. It maps such data into a different space where the data becomes linearly separable. The kernel function is defined as follows: </p><p>A disadvantage concerning this kernel is that is difficult to design, in the sense that it is difficult to obtain an optimum value for its parameter σ (sigma) and choose the corresponding C that works best for a given problem. The fact that certain combinations of σ and C make the SVM highly sensitive to training data also contributes to the error rate of the RBF-based SVM.</p><p>One of the advantages of the RBF kernel is that given the kernel, the weights, the number of support vector and the support vectors itself are automatically obtained as part of the training procedure, i.e. they don't need to be specified by the training mechanism.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>IV. SETTING THE EXPERIMENTS</head><p>As part of the experiments it was decided to work with two sets of images, one set consisting of coastal beach scenes, and the other set consisting of city scenes images. The processing of input images is done in order to reduce computational complexity.</p><p>The first set is conformed by 128 images of beach scenes content, the second set is conformed by 128 images of city scenes content, a total of 256 images.</p><p>The input images after processed end up being 8-bit per pixel grayscale images of dimension 128x128, working with just one channel reduces de number of convolutions. The output of each filter is obtained by the convolution of the input image with a Gabor filter. The process is shown below: </p><p>This computation can theoretically be done in the spatial domain however the Gabor filter is usually narrow. The filter is usually much larger in the frequency domain and thus less affected by aliasing effects due to sampling. It is thus more convenient to do all the computation process in the frequency domain. The convolution is then reduced to a simple and efficient point-wise multiplication of the Fourier transforms <ref type="bibr">[11]</ref>.</p><p>Iteratively, the sigma values were changed until a significant error rate is obtained. The test results for the learned algorithm are presented in table <ref type="table">II</ref>.</p><p>As it can be seen in table II the sigma value which represents the lower percentage error is σ = 35, with an error rate of 9.37%. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>VII. CONCLUSIONS</head><p>Extracting texture features by a Gabor filter bank and classify the filter outputs via the Support Vector Machines offers an excellent accuracy rate, 90.63% of the input images are correctly classified according to their class, belonging to beach scenes class or city scenes class. The article proves the efficiency of using a dual model, first to extract de texture features and then classify them with SVM.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 .</head><label>1</label><figDesc>Figure 1. "Traditional" Web Pages with hyperlinks</figDesc><graphic coords="6,394.95,322.36,162.00,101.25" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 .</head><label>2</label><figDesc>Figure 2. The Semantic Web Layers</figDesc><graphic coords="7,38.35,424.97,245.55,154.50" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 .</head><label>3</label><figDesc>Figure 3. Relationship between RDF and RDFS[5]    </figDesc><graphic coords="8,38.35,72.00,243.10,130.30" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 4 .</head><label>4</label><figDesc>Figure 4. A Revised Semantic Web Layer Cake</figDesc><graphic coords="9,41.06,129.50,236.19,235.97" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 5 .</head><label>5</label><figDesc>Figure 5. Web of Data (fragment) 12 [July 2009]</figDesc><graphic coords="10,46.65,443.62,223.94,201.52" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head></head><label></label><figDesc>description of the field An example could be the next one: | genre = Science fiction 2 Data was collected on February 2009.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. DIVEDD architecture.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_8"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. Corpus processing.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_9"><head>8Fig. 3 .</head><label>3</label><figDesc>Fig. 3. DIVEDD interface.</figDesc><graphic coords="28,144.84,147.36,305.92,197.88" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_10"><head>Fig. 4 .</head><label>4</label><figDesc>Fig. 4. Extraction of real examples of tener broncas.</figDesc><graphic coords="28,136.14,374.22,302.34,193.74" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_11"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Multidimensionality in dynamic taxonomies.</figDesc><graphic coords="33,143.15,147.35,309.00,141.95" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_12"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. Left side: Data set {a,b,…,j} classified under two facets. Right side: Reduced taxonomy.</figDesc><graphic coords="33,130.55,363.10,334.20,129.15" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_13"><head>Fig. 3 .</head><label>3</label><figDesc>Fig. 3. An artistic object with their textual description (denotation).</figDesc><graphic coords="34,155.65,147.35,284.15,127.20" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_14"><head>Fig. 5 .</head><label>5</label><figDesc>Fig. 5. Left side: Expanding the concepts. Right side: Concept expansion.</figDesc><graphic coords="35,150.55,491.16,294.40,118.54" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_15"><head></head><label></label><figDesc>Fig 7 we show a simple example of the taxonomy for the facets.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_16"><head>Fig. 7 .</head><label>7</label><figDesc>Fig. 7. A facet taxonomy example.</figDesc><graphic coords="36,128.30,440.40,338.85,103.55" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_17"><head>Fig. 8 .</head><label>8</label><figDesc>Fig. 8. Denotation index.</figDesc><graphic coords="37,175.39,231.89,244.71,144.31" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_18"><head>Fig. 9 .</head><label>9</label><figDesc>Fig. 9. Concepts connections.</figDesc><graphic coords="37,127.78,462.08,339.87,188.42" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_19"><head>Fig. 10 .</head><label>10</label><figDesc>Fig. 10. Conceptual index.</figDesc><graphic coords="38,176.40,147.35,242.65,141.95" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_20"><head>Fig. 11 .</head><label>11</label><figDesc>Fig. 11. EER Model for store a facet taxonomy and artworks objects.</figDesc><graphic coords="38,125.12,387.12,345.38,170.78" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_21"><head>Fig. 12</head><label>12</label><figDesc>Fig. 12 Recall and Precision evaluation.</figDesc><graphic coords="39,204.70,313.20,185.90,129.50" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_22"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. The new coordinates system used in the implemented hash-based function for fingerprinting.</figDesc><graphic coords="43,217.75,115.86,179.67,179.90" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_23"><head>4 F.</head><label>4</label><figDesc>Zacar ías et al.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_24"><head>6 F.</head><label>6</label><figDesc>Zacar ías et al.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_25"><head>Fig. 2 .</head><label>2</label><figDesc>Fig.2. Process for answer definition questions[7]    </figDesc><graphic coords="53,163.67,116.47,288.00,215.47" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_26"><head>Fig. 3 .</head><label>3</label><figDesc>Fig.3. Rga WAP Programming Model[9]    </figDesc><graphic coords="55,199.67,486.46,216.00,143.44" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_27"><head>10 F.</head><label>10</label><figDesc>Zacar ías et al.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_28"><head>Fig. 4 .</head><label>4</label><figDesc>Fig. 4. Main mQA Architecture</figDesc><graphic coords="57,217.68,116.05,180.00,143.88" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_29"><head></head><label></label><figDesc>(the nodes have dual microprocessors); we would have 50 bamboo nodes in each node ROL having a 25/CPU relation. This script calls the name.cfg configuration file and this also calls the DHT class and semantic class.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_30"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. Comparison of Routings.</figDesc><graphic coords="69,158.50,451.90,280.10,161.30" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_31"><head>Fig 1 .</head><label>1</label><figDesc>Fig 1. System Architecture</figDesc><graphic coords="73,184.50,147.42,237.66,176.10" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_33"><head>Fig 2 .</head><label>2</label><figDesc>Fig 2. Services Registration Architecture</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_34"><head>Fig 3 .</head><label>3</label><figDesc>Fig 3. Database Relations Diagram.</figDesc><graphic coords="74,158.82,452.52,289.16,230.28" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_35"><head>Fig 4 .</head><label>4</label><figDesc>Fig 4. Building a User Profile Architecture</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_36"><head>Fig. 1</head><label>1</label><figDesc>Fig.1 Representation of the Iris Data SetFor the experimental tests carried, the following data sets have been used: Iris, Glass and Wine. This report presents the results for the Iris data set.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_38"><head>Fig. 2 %</head><label>2</label><figDesc>Fig. 2 % of Exchanged Points.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_39"><head>Fig. 2</head><label>2</label><figDesc>Fig. 2 corresponds to the test I4 and the graphical representation of the exchange behavior at each iteration. Likewise Fig. 3 represents the behavior of the sum of distances for the same test.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_41"><head>Fig. 3</head><label>3</label><figDesc>Fig. 3 Total Sum of Distance No. Prue. No. Iter. No. Ptos.Interc.2 % Dif. Ptos. Interc. 1 a 2</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_44"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Using the Kernel to transform (map) the input data space.</figDesc><graphic coords="105,179.04,571.86,249.18,95.16" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0"><head></head><label></label><figDesc></figDesc><graphic coords="1,87.00,70.85,442.20,216.00" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0"><head></head><label></label><figDesc></figDesc><graphic coords="52,163.67,116.14,288.00,251.79" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0"><head></head><label></label><figDesc></figDesc><graphic coords="109,173.64,181.92,247.92,210.00" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head></head><label></label><figDesc>Scientific American article has not yet arrived. But perhaps it is arriving by stealth, under the guise of the "Web 3.0" umbrella. Confusion still abounds about the meaning of the term "Web 3.0", which has been variously described as being about the meaning of data, intelligent search, or a "personal assistant". This sounds like what the Semantic Web has to offer, but even if the terms do not become synonymous, it is clear that the Semantic Web will form a crucial component of Web 3.0 (or vice versa!).</figDesc><table /><note>11 http://www.ted.com/talks/tim_berners_lee_on_the_next_web.html</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>[ 5 ]</head><label>5</label><figDesc>Fensel D, Hendler JA, Lieberman H, WahlsterW(eds). 2003. Spinning the Semantic Web: Bringing the World Wide Web to its Full Potential. MIT Press: Cambridge, MA. ISBN 0-262-06232-1. [6] Gruber, T. 1993. Toward principles for the design of ontologies used for knowledge sharing. In Guarino N, Poli R (eds). International Workshop on Formal Ontology, Padova, Italy, [7] Hendler, J., 2008. Linked Data: The Dark Side of the Semantic Web,</figDesc><table><row><cell cols="8">(tutorial), 7th International Semantic Web Conference (ISWC08),</cell></row><row><cell cols="2">Karlsruhe, Germany.</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>[8] Linking</cell><cell>Open</cell><cell></cell><cell>Data</cell><cell>Wiki,</cell><cell></cell><cell>available</cell><cell>at</cell></row><row><cell cols="8">http://esw.w3.org/topic/SweoIG/TaskForces/CommunityProjects/Lin-</cell></row><row><cell>kingOpenData</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="8">[9] Manning, C., Schütze, H., 1999. Foundations of statistical natural</cell></row><row><cell cols="4">language processing. MIT Press.</cell><cell></cell><cell></cell><cell></cell></row><row><cell>[10] "Semantic</cell><cell>Web</cell><cell>-</cell><cell cols="2">XML2000,</cell><cell>slide</cell><cell>10".</cell><cell>W3C.</cell></row><row><cell cols="7">http://www.w3.org/2000/Talks/1206-xml2k-tbl/slide10-0.html.</cell></row><row><cell cols="8">[11] Shadbolt, N., Hall, W., Berners-Lee, T., 2006. The Semantic Web</cell></row><row><cell>Revisited.</cell><cell></cell><cell>IEEE</cell><cell></cell><cell cols="2">Intelligent</cell><cell cols="2">Systems.</cell></row><row><cell cols="8">http://eprints.ecs.soton.ac.uk/12614/1/Semantic_Web_Revisted.pdf.</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_6"><head>Numericaldatafoundovertheanalysisofinfoboxes</head><label></label><figDesc></figDesc><table><row><cell>Field name</cell><cell>Number of full fields</cell><cell>Number of empty fields</cell></row><row><cell>Director</cell><cell>5,092</cell><cell>0</cell></row><row><cell>Título</cell><cell>5,092</cell><cell>0</cell></row><row><cell>Título ID</cell><cell>5,092</cell><cell>0</cell></row><row><cell>Título orignal</cell><cell>4,897</cell><cell>195</cell></row><row><cell>Año</cell><cell>5,069</cell><cell>23</cell></row><row><cell>Género</cell><cell>4,499</cell><cell>593</cell></row><row><cell>Director</cell><cell>5,092</cell><cell>0</cell></row><row><cell>Título</cell><cell>5,092</cell><cell>0</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_8"><head>4 Arturo Velasco, María J. Somodevilla, and Ivo. H. PinedaTable 1 .</head><label>1</label><figDesc>Handling synonym expressions in the DIVEDD.</figDesc><table><row><cell></cell><cell>+Frequent_ Verbal_ Expression</cell><cell>Definition</cell><cell>Key-word</cell><cell>Thematic_ Field</cell><cell>Linguistic_ Record</cell><cell>Country</cell></row><row><cell>S y n o n</cell><cell>Ir al bote Ir al tambo</cell><cell>Meter a alguien en la cárcel Meter a alguien en la cárcel</cell><cell>Bote Tambo</cell><cell>Behavior Behavior</cell><cell>Informal Informal</cell><cell>México México</cell></row><row><cell>y m s</cell><cell>Ir a la sombra</cell><cell>Meter a alguien en la cárcel</cell><cell>Sombra</cell><cell>Behavior</cell><cell>Informal</cell><cell>México</cell></row><row><cell></cell><cell>Ir tras las rejas</cell><cell>Meter a alguien en la cárcel</cell><cell>Rejas</cell><cell>Behavior</cell><cell>Informal</cell><cell>México</cell></row><row><cell></cell><cell>Hacer la barba</cell><cell>Lisonjear a alguien</cell><cell>Barba</cell><cell>Behavior</cell><cell>Informal</cell><cell>México</cell></row><row><cell></cell><cell>Hacer la pelota</cell><cell>Lisonjear a alguien</cell><cell>Pelota</cell><cell>Behavior</cell><cell>Informal</cell><cell>Spain</cell></row><row><cell></cell><cell>Hacer la rosca</cell><cell>Lisonjear a alguien</cell><cell>Rosca</cell><cell>Behavior</cell><cell>Informal</cell><cell>Spain</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_9"><head></head><label></label><figDesc>are described in table 4 in section 3.4, Generation of Synonyms and Variants.</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_10"><head>Table 2 .</head><label>2</label><figDesc>Variants expressions in the database of the DIVEDD.</figDesc><table><row><cell>+Frequent_</cell><cell></cell><cell></cell><cell></cell></row><row><cell>Verbal_</cell><cell>Definition</cell><cell>Key-word</cell><cell>Variants_Verbal_Expression</cell></row><row><cell>Expression</cell><cell></cell><cell></cell><cell></cell></row><row><cell>Ir al bote</cell><cell>Meter a alguien en la cárcel</cell><cell>Bote</cell><cell>[ir,llevar,meter] (al) {bote} [refundir] (en_el) {bote}</cell></row><row><cell>Ir al tambo</cell><cell>Meter a alguien en la cárcel</cell><cell>Tambo</cell><cell>[ir,llevar,meter] (al) {tambo} [refundir] (en_el) {tambo}</cell></row><row><cell>Ir a la sombra</cell><cell>Meter a alguien en la cárcel</cell><cell>Sombra</cell><cell>[ir,llevar,meter] (a_la) {sombra} [refundir] (en_la) {sombra}</cell></row><row><cell>Ir tras las rejas</cell><cell>Meter a alguien en la cárcel</cell><cell>Rejas</cell><cell>[ir,llevar,meter,refundir] (tras_las) {rejas}</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_11"><head></head><label></label><figDesc>Field of the tableVariant that stores the variants of the canonical verbal expressions.</figDesc><table><row><cell>Variant_</cell></row><row><cell>Verbal_</cell></row><row><cell>Expression</cell></row></table><note>Country of origin of the expression. Region area or region of use of the expression. Thematic_ Field Thematic field of the expression Key-Word Alexical component unit of the expression. Useful to distinguish among synonym expressions. Example Field of the table Examples that stores the examples provided by the lexicographer and extracted of the corpus.</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_12"><head>Table 1 .</head><label>1</label><figDesc>Evaluation of the WePS-2 test dataset.</figDesc><table><row><cell>FMeasure 0.2 FMeasure 0.2 FMeasure 0.5 FMeasure 0.5</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_13"><head>Table 2 .</head><label>2</label><figDesc>Experimental results with the training dataset of the WePS-1 collection.</figDesc><table><row><cell>FMeasure 0.5 FMeasure 0.5</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_14"><head>Table 3 .</head><label>3</label><figDesc>Experimental results with the test dataset of the WePS-1 collection.</figDesc><table><row><cell>FMeasure 0.5 FMeasure 0.5</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_15"><head></head><label></label><figDesc>Bamboo, name_file. cfg. Each node of the bamboo network must have a fill of these characteristics</figDesc><table><row><cell></cell><cell cols="3">class bamboo.dmgr.DataManager</cell></row><row><cell cols="4">&lt;initargs&gt; #Routing dht_bamboo</cell></row><row><cell></cell><cell cols="3">debug_level &lt;StageSemantic_dht&gt;</cell><cell>0</cell></row><row><cell></cell><cell cols="3">merkle_tree_expansion 2 class bamboo.cont_dht.StageSemantic_dht</cell></row><row><cell></cell><cell cols="3">desired_replicas &lt;initargs&gt;</cell><cell>2</cell></row><row><cell></cell><cell cols="3">&lt;/initargs&gt; debug_level 1</cell></row><row><cell></cell><cell cols="3">&lt;/DataManager&gt; &lt;/initargs&gt;</cell></row><row><cell></cell><cell cols="3">&lt;/StageSemantic_dht&gt;</cell></row><row><cell cols="4">&lt;sandstorm&gt; &lt;StorageManager&gt;</cell></row><row><cell></cell><cell cols="3">&lt;global&gt; class bamboo.db.StorageManager &lt;/stages&gt;</cell></row><row><cell cols="4">&lt;initargs&gt; &lt;initargs&gt; &lt;/sandstorm&gt;</cell></row><row><cell>#</cell><cell cols="3">node_id ${NodeID} debug_level</cell><cell>0</cell></row><row><cell></cell><cell cols="3">&lt;/initargs&gt; homedir</cell><cell>${CacheDir}</cell></row><row><cell></cell><cell cols="3">&lt;/global&gt; &lt;/initargs&gt;</cell></row><row><cell></cell><cell cols="3">&lt;/StorageManager&gt;</cell></row><row><cell></cell><cell>&lt;stages&gt;</cell><cell></cell></row><row><cell></cell><cell cols="3">&lt;DataManagerTest&gt;</cell></row><row><cell></cell><cell cols="3">&lt;Network&gt; class bamboo.dmgr.DataManagerTest</cell></row><row><cell></cell><cell cols="3">class bamboo.lss.Network &lt;initargs&gt;</cell></row><row><cell>#</cell><cell cols="3">&lt;initargs&gt; debug_level</cell><cell>0</cell></row><row><cell>#</cell><cell></cell><cell></cell><cell>udpcc_debug_level 0 to_put 0</cell></row><row><cell></cell><cell></cell><cell></cell><cell>drop_prob 0.1 put_size</cell><cell>100</cell></row><row><cell></cell><cell cols="3">&lt;/initargs&gt; &lt;/initargs&gt;</cell></row><row><cell></cell><cell cols="3">&lt;/Network&gt; &lt;/DataManagerTest&gt;</cell></row><row><cell></cell><cell cols="2">&lt;Rpc&gt; &lt;Dht&gt;</cell></row><row><cell></cell><cell cols="3">class bamboo.lss.Rpc class bamboo.dht.Dht</cell></row><row><cell></cell><cell cols="3">&lt;initargs&gt; &lt;initargs&gt;</cell></row><row><cell></cell><cell cols="3">&lt;/initargs&gt; #debug_level 1</cell></row><row><cell></cell><cell>&lt;/Rpc&gt;</cell><cell></cell><cell>storage_manager_stage StorageManager</cell></row><row><cell></cell><cell></cell><cell></cell><cell>min_replica_count</cell><cell>1</cell></row><row><cell></cell><cell>&lt;Router&gt;</cell><cell cols="2">&lt;/initargs&gt;</cell></row><row><cell></cell><cell cols="3">class bamboo.router.Router &lt;/Dht&gt;</cell></row><row><cell></cell><cell cols="3">&lt;initargs&gt;</cell></row><row><cell>#</cell><cell cols="3">debug_level &lt;Gateway&gt;</cell><cell>1</cell></row><row><cell></cell><cell cols="3">gateway class bamboo.dht.Gateway ${GatewayID}</cell></row><row><cell></cell><cell cols="3">periodic_ping_period &lt;initargs&gt;</cell><cell>20</cell></row><row><cell></cell><cell cols="3">ls_alarm_period # debug_level 1</cell><cell>4</cell></row><row><cell></cell><cell cols="3">near_rt_alarm_period port ${GatewayPort} 0</cell></row><row><cell></cell><cell cols="3">far_rt_alarm_period &lt;/initargs&gt;</cell><cell>10</cell></row><row><cell>#</cell><cell cols="3">leaf_set_size &lt;/Gateway&gt;</cell><cell>8</cell></row><row><cell></cell><cell cols="3">digit_values</cell><cell>2</cell></row><row><cell></cell><cell cols="3">ignore_proximity &lt;WebInterface&gt;</cell><cell>false</cell></row><row><cell></cell><cell cols="3">location_cache_size class bamboo.www.WebInterface 0</cell></row><row><cell></cell><cell cols="3">&lt;/initargs&gt; &lt;initargs&gt;</cell></row><row><cell></cell><cell cols="2">&lt;/Router&gt;</cell><cell>storage_manager_stage StorageManager</cell></row><row><cell></cell><cell cols="3">&lt;/initargs&gt;</cell></row><row><cell></cell><cell cols="3">&lt;DataManager&gt; &lt;/WebInterface&gt;</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_16"><head></head><label></label><figDesc>1. Types of messages. 2. The payload or what is the message. 3. The different logs files. Format of Logs. The logs stay in each node bamboo: Log of requests that indicates the objects that stay in the node. Log of number of hops to reach an object. It registers the object number and the number of hops that it required to arrive to the object.</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_17"><head>Table 1 .</head><label>1</label><figDesc>Bamboo-DHT Routing parameters.</figDesc><table><row><cell>Scenes:</cell><cell>No. Nodes</cell><cell>No.</cell><cell>No. Requests</cell><cell>Without caching,</cell></row><row><cell>Bamboo-</cell><cell>(parameter 2)</cell><cell>Objects</cell><cell>(parameter 1)</cell><cell>without metadata</cell></row><row><cell>DHT</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>1</cell><cell>500</cell><cell>100</cell><cell>10000</cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_18"><head>Table 2 .</head><label>2</label><figDesc>Semantic Routing parameters.</figDesc><table><row><cell>Scenes:</cell><cell>Nodes</cell><cell>Objects</cell><cell>Requests</cell><cell>Size</cell><cell>of</cell></row><row><cell>info-semantic</cell><cell>(parameter 2)</cell><cell></cell><cell>(parameter 1)</cell><cell>caching</cell><cell></cell></row><row><cell>(parameter 4)</cell><cell></cell><cell></cell><cell></cell><cell cols="2">(parameter 3)</cell></row><row><cell>1</cell><cell>500</cell><cell>100</cell><cell>10000</cell><cell>5</cell><cell></cell></row><row><cell>2</cell><cell>500</cell><cell>250</cell><cell>12500</cell><cell>5</cell><cell></cell></row><row><cell>3</cell><cell>500</cell><cell>500</cell><cell>15000</cell><cell>5</cell><cell></cell></row><row><cell>4</cell><cell>500</cell><cell>750</cell><cell>15000</cell><cell>10</cell><cell></cell></row><row><cell>5</cell><cell>500</cell><cell>1000</cell><cell>20000</cell><cell>15</cell><cell></cell></row><row><cell>6</cell><cell>500</cell><cell>1000</cell><cell>20000</cell><cell>20</cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_19"><head>Table</head><label></label><figDesc>Are is connected to table Recom rvices with a recommendation list. Its primary key references to SID (services) and RID (recommendation).The last table is Rec imary key references to UUID (users) and RID (recommendation).</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_21"><head>•</head><label></label><figDesc>Structure (syntax of the compound sentence, what relationships are established: causeconsequence, principlepurpose, conditional potential, hypothesis, etc.); the ideas exposition and their linking; structure consistent with the basic genre on introduction (definition, description, narrative, argumentation), development and conclusion. • Functional: the event affects the purpose, theme (or its treatment), target and record/register (what, for whom, for what purpose, how); informative function, logic, data or knowledge transfer, assertive or speculative, formalized as article, essay, notes, summary; without subjective, intimate or emotional levels. • Style: neutral and objective tone; author plural or third impersonal; without dialect marks; rigorous, light, clear and simple.</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_22"><head></head><label></label><figDesc>Answer outside of the semantic field. Unbelievable at firs glance, among the highest percentages it was found the term vaya (exclamation and verb) with the 77,77%, followed by haciendo, with 69,62%, vario, with 57,77%. Meanings or homonymy omisión (we are only interested on those answers that show significant plurality but not the classification of the term meaning of homonym term). This is considered "partial correct answer". In rebelar, it was obtained a 41,1%; in revelar the 46,6%; in ascender the 48,8%, i.e., the majority. 10. Figurative sense, without the denotative. árido: 'falto de amenidad' 11. Semantic shift. árido: 'molesto' 12. Subsystem of use or colloquial. vacilar: 'echar relajo' 13. Answer without meaning. arriar: 'arroyar' [ por arrollar]</figDesc><table><row><cell cols="3">of logicalgrammatical category within the term meaning. Simple. Term: soluble (adj.) Definition given: 'disolver' (verb); Term: cesión (noun) Definition given: 'que se termina o sesa 3. Confusion with its homophones (diacritical spelling) or paronymie. The value 1. • • [sic]; ranges from 0% for vaso (basobazo) to 17,7%, 17,3% and 14,8% for asechar, adoleciente, rebelar, which are defined by their pair acechar, adolescente, revelar, respectively. 1. Homophone category error. acerbo (adj.): conjunto de bienes comunes (corresponde a acervo, sust.) 2. Outside of its category [Example. Acerbo (adj.): tener conocimientos (they mislead with acervo, sust., but they defined it as verb) 3. Half confusion (one category inside and one category outside). acerbo: 'acumular, conjunto de cosas' 4. Examples of wrong field: • Term: adolescente [person] Definition given: 'etapa'; 'edad' [cosa]; • Term: asciendo (present tense verb, act) Definition given: 'posición en que uno queda' [resultado de un acto]; • Term: revelar, Definition given: 'acción quimica [sic] de la fotografía' 5. Definition by means of: • Temporality: es cuando (ej. estática: cuando una persona no se mueve) • Mode: es como...(it is like…) (ex. occisa: es como estar muerta) • Location: es donde... (is where…) (ex. jardín: donde juegan los niños) • Function: sirve para... (it is to…) (ex. lenguaje: sirve para comunicarse) 6. To classify instead of defining. Very often they use to annotate the infinite without definition (nor meaning) in the particular case of conjugated verbs. Example. izo: verbo izar; vacilo, 'de vacilar'. 7. To complete instead of defining. (example: acervo: bibliográfico; sucesión: presidencial) 8. The efinition was to much… • Extended, wide (ex. es una cosa...; es algo...) • Restricted (ex. acervo: conjunto de palabras) • Fuzzy, imprecise (ex. bacilo: pertenece a la biología) type. • Acervo:: 'vacteria' [sic]; 'es la consistencia de algunas cosas, p.e. la piña'; 'afirmar'; 'contar algo'; 'hacer muchas cosas' • Revelar: 'identificar caracteres'; 'presentar algo a escala'; 'quitar'; 'relativo a copiar, algo oculto'; 'descifrar'. • Vario: 'del verbo vasar [sic]'; 'alguna cosa de uso común'; 'mucho' 9. In summary, below we show some other examples without classifying the specific error • Vacilo 'que no tiene seguridad'; 'indeciso'</cell></row><row><cell>•</cell><cell>Term: suspendido</cell><cell>Definition given: 'cuando algo a [sic]</cell></row><row><cell>2.</cell><cell cols="2">terminado ó [sic] concluido' Half confusion (one category inside and one category outside).</cell></row><row><cell>•</cell><cell>Term: acceso (sust.)</cell><cell>Definition given: entrada (sust.), viable</cell></row><row><cell></cell><cell>(adj.);</cell><cell></cell></row><row><cell>•</cell><cell>Term: sesión (sust.) (verbo), reunión (sust.)'</cell><cell>Definition given: 'congregarse personas</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_23"><head></head><label></label><figDesc>Centro Nacional de Investigación y Desarrollo Tecnológico, Cuernavaca Mor. Mex. 2 Benemérita Universidad Autónoma Puebla, Fac. Cs. de la Computación, México. jperez@cenidet.edu.mx,{rboone,mariasg}@cs.buap.mx Abstract. Clustering problems arise in many different applications: machine learning data mining and knowledge discovery, data compression and vector quantization, pattern recognition and pattern classification. It is considered that the k-means algorithm is the best-known squared errorbased clustering algorithm, is very simple and can be easily implemented in solving many practical problems.</figDesc><table><row><cell>Algorithm: An Experimental</cell></row><row><cell>Trial Using Matlab</cell></row><row><cell>Joaquín Pérez Ortega 1 , Ma. Del Rocío Boone Rojas, 1,2 , María J. Somodevilla</cell></row><row><cell>García 2</cell></row><row><cell>1</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_24"><head></head><label></label><figDesc>includes the Iris data, for illustrative purposes only are considered the attributes sepal length, petal length and petal width.</figDesc><table><row><cell cols="5">Test I4: &gt;&gt; [u,v,sumd,D]= kmeans(z,3,'display','iter');</cell></row><row><cell>iter</cell><cell cols="2">phase num</cell><cell>sum</cell><cell>%inter</cell></row><row><cell>1</cell><cell>1</cell><cell>150</cell><cell cols="2">426.888 100</cell></row><row><cell>2</cell><cell>1</cell><cell>34</cell><cell cols="2">134.187 22.6</cell></row><row><cell>3</cell><cell>1</cell><cell>13</cell><cell cols="2">105.771 8.6</cell></row><row><cell>4</cell><cell>1</cell><cell>12</cell><cell cols="2">88.8948 8</cell></row><row><cell>5</cell><cell>1</cell><cell>6</cell><cell cols="2">85.2326 4</cell></row><row><cell>6</cell><cell>1</cell><cell>4</cell><cell cols="2">84.064 2.6</cell></row><row><cell>7</cell><cell>1</cell><cell>3</cell><cell cols="2">83.3704 2</cell></row><row><cell>8</cell><cell>1</cell><cell>5</cell><cell cols="2">82.073 3.3</cell></row><row><cell>9</cell><cell>1</cell><cell>3</cell><cell cols="2">81.3672 2</cell></row><row><cell>10</cell><cell>1</cell><cell>4</cell><cell cols="2">80.3157 2.6</cell></row><row><cell>11</cell><cell>1</cell><cell>3</cell><cell cols="2">79.6817 2</cell></row><row><cell>12</cell><cell>1</cell><cell>3</cell><cell cols="2">79.1156 2</cell></row><row><cell>13</cell><cell>1</cell><cell>1</cell><cell cols="2">78.9451 0.6</cell></row><row><cell>14</cell><cell>2</cell><cell>1</cell><cell cols="2">78.9408 0.6</cell></row><row><cell cols="5">14 iterations, total sum of distances = 78.9408</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_25"><head>Table 1</head><label>1</label><figDesc>Summary of Results for Iris Data Set 92 5.1 Summary of Results for the Iris data set.Table1provides a summary of 25 experimental tests performed on the IRIS data set, the first column identifies the test number and the second column includes the number of iterations performed by the algorithm.</figDesc><table><row><cell>1</cell><cell>5</cell><cell>14</cell><cell>90.7</cell></row><row><cell>2</cell><cell>10</cell><cell>7</cell><cell>95.4</cell></row><row><cell>3</cell><cell>3</cell><cell>5</cell><cell>96.7</cell></row><row><cell>4</cell><cell>6</cell><cell>8</cell><cell>94.7</cell></row><row><cell>5</cell><cell>14</cell><cell>34</cell><cell>77.4</cell></row><row><cell>6</cell><cell>4</cell><cell>5</cell><cell>96.7</cell></row><row><cell>7</cell><cell>8</cell><cell>8</cell><cell>94.7</cell></row><row><cell>8</cell><cell>8</cell><cell>36</cell><cell>97.6</cell></row><row><cell>9</cell><cell>7</cell><cell>73</cell><cell>95.4</cell></row><row><cell>10</cell><cell>3</cell><cell>4</cell><cell>97.4</cell></row><row><cell>11</cell><cell>7</cell><cell>37</cell><cell>97.6</cell></row><row><cell>12</cell><cell>11</cell><cell>9</cell><cell>94.0</cell></row><row><cell>13</cell><cell>6</cell><cell>53</cell><cell>64.7</cell></row><row><cell>14</cell><cell>5</cell><cell>12</cell><cell>92.0</cell></row><row><cell>15</cell><cell>3</cell><cell>2</cell><cell>98.7</cell></row><row><cell>16</cell><cell>6</cell><cell>15</cell><cell>90.0</cell></row><row><cell>17</cell><cell>4</cell><cell>3</cell><cell>98.0</cell></row><row><cell>18</cell><cell>10</cell><cell>9</cell><cell>96.0</cell></row><row><cell>19</cell><cell>10</cell><cell>4</cell><cell>97.4</cell></row><row><cell>20</cell><cell>6</cell><cell>32</cell><cell>79.0</cell></row><row><cell>21</cell><cell>12</cell><cell>16</cell><cell>89.4</cell></row><row><cell>22</cell><cell>8</cell><cell>33</cell><cell>78.0</cell></row><row><cell>23</cell><cell>7</cell><cell>31</cell><cell>80.0</cell></row><row><cell>24</cell><cell>6</cell><cell>14</cell><cell>90.7</cell></row><row><cell>25</cell><cell>11</cell><cell>7</cell><cell>95.4</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_26"><head>Table 1 :</head><label>1</label><figDesc>It can see that for 150 points in the Iris database, and a set of 25 tests, the k-means algorithm in Matlab: Converge in an average of 7.2 iterations. The average number of points exchanged during the second iteration was 18.84 points The percentage of points located on the second iteration in the corresponding group was 91.0%The results of the analysis for our sample work, allow us to establish a framework and analyze the theoretical study of the k-means algorithm Also we reseach and distinguish the different lines on which there is still a fertile field for investigation. As we can see several attempts at overcoming the shortcomings of the k-means algorithm have been done and different approaches in different disciplines have been proposed: Optimization, Probability and Statistics, Neural Networks, Evolutionary Algorithms, among others. The vast majority of contributions have focused on the first three lines of research identified in this study: The sensibility of the algorithm to initial conditions, convergence of the algorithm to a local optimum rather than a global optimum and the efficiency of the algorithm. Notes that challenges still to be resolved in such research and has been relatively little work done on the lines related to the implementation of the algorithm to other variables as well as treatment to outliers and noise. Acording the tests conducted in Matlab, this laboratory showed that it is actually very conducive to experimental testing, the implementation of k-means, allows to monitor the performance of the algorithm through the information that can be deployed at runtime, such as result of the objective function and the number of points exchanged in each iteration.The results allow us to establish a framework to compare the proposal improvement algorithm with the previous work. As part of this project and to give continuity to previous work[29] [30], also ventures into different applications to k-means, such as in the areas of health care in Mexico and in Web Usage Mining for Log files from the server of the Faculty of Compute Science BUAP, México.</figDesc><table><row><cell>6 Conclusions.</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_27"><head>Table 2 .</head><label>2</label><figDesc>Experimental results for different sigma values</figDesc><table><row><cell>RBF</cell><cell>error</cell><cell>RBF</cell><cell>error</cell></row><row><cell>σ=21</cell><cell>0.1406</cell><cell>σ=29</cell><cell>0.1094</cell></row><row><cell>σ=22</cell><cell>0.1094</cell><cell>σ=30</cell><cell>0.1094</cell></row><row><cell>σ=23</cell><cell>0.1094</cell><cell>σ=31</cell><cell>0.1094</cell></row><row><cell>σ=24</cell><cell>0.1094</cell><cell>σ=32</cell><cell>0.1094</cell></row><row><cell>σ=25</cell><cell>0.1094</cell><cell>σ=33</cell><cell>0.1094</cell></row><row><cell>σ=26</cell><cell>0.1094</cell><cell>σ=34</cell><cell>0.1094</cell></row><row><cell>σ=27</cell><cell>0.1094</cell><cell>σ=35</cell><cell>0.09375</cell></row><row><cell>σ=28</cell><cell>0.1094</cell><cell>σ=36</cell><cell>0.09375</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_0">http://www.w3.org/TR/rdf-schema/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_1">http://sourceforge.net/projects/threestore/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_2">http://jena.sourceforge.net/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="10" xml:id="foot_3">www.twine.com</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_4">http://en.wikipedia.org/wiki/Academic_Research_on_Wikipedia.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_5">Snowball, Spanish stop word list, http://snowball.tartarus.org/algorithms/spanish/stop.txt</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_6">Ranks NL, Spanish stopwords, http://www.ranks.nl/stopwords/spanish.html</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_7">Snowball, Spanish stemming algorithm, http://snowball.tartarus.org/algorithms/spanish/stemmer.html.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_8">Information Management Unit, Guided Interactive Discovery of e-Government Services, http://www.imu.iccs.gr/sweg/presentations/Giovanni%20Maria%20Sacco.ppt</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" xml:id="foot_9">⋆ This work has been partially supported by the CONACYT project #106625, as well as by the PROMEP/103.5/09/4213 grant.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_10">There is a theoretical discussion about the status of the studies dealing with the language, from grammar to the diverse branches, lines and research. By centuries it was considered the grammar and rhetoric and oratory as complete arts. However for the end of last century, with F. de Saussure as a benchmark, it was understood as a science. In this paper we will no go deeper in order to argue about one or other viewpoint.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_11">We would argue whether or not the natural language is the knowledge instrument or the knowledge itself with respect to its role as reality creator, in terms of the rational speech or thought. How do we know? How do we learn the reality? By means of which mechanism or procedure we "know" and "know the world".</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_12">In both senses, strict and soft, i.e., we include here the meaning plurality.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_13">The term 'in general' is due to some signs (verbs) where the word is the significant thing. In other words, where decir (to say) is hacer (to do), like to judge, to promise in contrast with comer (to eat), correr (to run), etc.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_14">The intent separates symbol or sign or symptom or sign of evidence, not intentional. The same fact can be used as a sign or signal. Let us take for instance the Vatican smoke and that one in a forest, they both 'indicate' that there is fire, but the former is deliberate, intentional (it means that there are or not a new Pope, white or black, respectively); in the latter case, it indicates that there is fire.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_15">We exclude Saussure and Frege, because we consider them to be non functional. We refer to Ogden and Richards, whom are generally accepted in our environment.</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This research was made possible by the financial support of CONACYT (82050) and DGAPA-PAPIIT (IN403108). The authors wish to thank Sarahi Abrego Romero for the proofreading of this paper. Acknowledgments Thank you very much to the Autonomous University of Puebla for their financial support. This work was supported under project VIEP register number 15968. Also, we thank the support of the academic body: Sistemas de Informacin.</p></div>
			</div>

			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The service dislikes 2</p><p>The services dislikes a little 3</p><p>The services likes a little 4</p><p>The service likes 5</p><p>The service likes very much</p><p>We can note that, according to the rules for rating services, the rating number can be</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4">Learning rules</head><p>A first approach to learning rules has been developed. The rules try to solve the If a topic rating has not been modified in X period of time 0. lse odified in X period of time a negative value. This way we can now know that a user doesn't like certain topic.</p><p>posterity problem; this is for topics that the user rates as a like a time ago, but now that topic has no more interest for the user. For example: a year ago a user interacted with topics about Olympic games, and that topic has a rate of 1.5, now the user is no more interested in Olympic games, but the recommender agent continues recommending services related to that topic because of its rating. There is not enough (and also not advisable) that the user rate several times the service recommended with 0, in order to decrease its value. We can apply some rules that, given a period of time, analyses if the user has been interested in a topic. If the topic has not been modified in that period of time we can "reset" the value of the topic setting it on a value of 0.</p><p>Then if its value is negative Then do nothing Else reset its value to E //a topic rating has been m if it has been modified only decreasing it Then reset its value to 0. Else do nothing. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Authors Title and Commentary</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>3.2.5</head><p>The definition of "means" limits the application only to numerical variables.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Several works have been developed that extend the application of k-means for categorical variables or others:</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Authors</head><p>Títle and Commentary <ref type="bibr" target="#b124">[38]</ref> Song, Wei, Li Cheng Hua, Park, Soon Cheo. "Genetic Algorithm for text clustering using ontology and evaluating the vality of various semantic simility measures." Improving the kmeans algorithm by using a genetic algorithm that finds similarities conceptual. Based on ontology, thesaurus corpus for clustering of text fields.</p><p>[14] S. Gupata, K. Rao, &amp;Bhatnagar "K-means clustering algorithm for categorical attributes". Title explicit.</p><p>[16] Z. Huang.</p><p>"Extensions to the k-means algorithm for clustering large data sets with categorical values." Title explicit. <ref type="bibr">4</ref> The Algorithm k-means on Matlab.</p><p>Experimental tests were conducted for K-means in the Matlab <ref type="bibr">[25]</ref>. The Matlab (Matrix Laboratory) is both, an environment and programming language for numerical calculations with vectors and matrices. It is a product of the company The </p><p>The family of Gabor filters selected to set up the filter bank for the experiments in the frequency domain are: </p><p>A filters bank is constructed by changing values on the four parameters mentioned before, (f,θ, γ, η). For our experiments we only change of the parameters f and θ , which are the central frequency of the filter and the rotation angle of the filter, the other two parameters (γ and η) are set to be constant.</p><p>For the experiments it has been set up the bank with 2 frequencies and 3 orientations, to obtain a total of 6 filter outputs for each image on the respective set.</p><p>The selected parameters for the filter banks are: central frequency f = 0.1725, which outputs the two frequencies 0.1725, 0.1220. Three orientations θ = 0°, 60° and 120°, sharpness along the mayor axis γ=0.5, sharpness along the minor axis η=0.5.</p><p>In order to apply the filters bank to the sets of images, each image needs to be transform into frequency domain via the Fourier transform. The 2-D Fourier transform used for images can be defined as:</p><p>where ( , ) is the original input image</p><p>The convolution of the input image with the Gabor filter is performed. In domain frequency the convolution is represented in a point-to-point multiplication of the transformed image with the Gabor filter. </p><p>Once the filter output is obtained, G(x,y) needs to be transformed back to its spatial representation using the Inverse Fourier Transform in 2-D. <ref type="bibr">)</ref> is the image in frequency domain</p><p>After the transformation, normalization is applied to the output image in order to avoid effects by illumination.</p><p>At the end of normalization, we have a certain number of square matrices per filtered image; each matrix dimension is 128x128. The number of square matrices depends on the number of orientations and frequencies concerning the filters bank, in our experiments the filter bank consist of 2 frequencies and 3 orientations, so the number of output matrices is 6.</p><p>When convolution is performed some results are not useful especially if the image does not contain textures that respond meaningfully to the filter selected parameters. To reduce the problem all the outputs obtained by convolution are summed up to remove the results that are not relevant and to enhance those that helps to detect texture regions; this also helps to reduce dimensionality of the feature space by having one square matrix as an output, same size of the input image.</p><p>At this point we have one matrix per input image, reducing the dimensionality of input data. Each matrix is used to build up a feature matrix, which is going to serve as an input of the SVM classifier.</p><p>To complete the convolution of the input image with each one of the Gabor filters we take only the real part of the output filtered image. As mentioned in <ref type="bibr">[31]</ref>, by this way we can keep most the texture response information ignoring phase information.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Re( ( , ))</head><p>G x y <ref type="bibr">(12)</ref> Then we modified each output matrix of dimension 128x128 to construct the feature matrix. We take each matrix and transform it in a 1x16384 vector, each vector is then piled up with the next transformed matrix to form the feature matrix.</p><p>Finally we have a feature matrix of dimension 256x16384 which serves as input to the classifier.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>V. SVM CLASSIFIER</head><p>The goal of experimentation is to obtain a training model through SVM, which can be capable of separate a set of input images. Once we have the feature matrix with processed and filtered images we proceed with the SVM classification procedure. According to the nature of the classification process we need to define a training dataset, so the classifier could learn a model, and a test dataset, that let us test the learned model. The training dataset is conformed by 75% of the input dataset, and the test dataset by the remaining 25%. The selection criteria to build up the training dataset and the test dataset are done randomly. In fig. <ref type="figure">2</ref>  The experiments were performed using SPIDER <ref type="bibr">[32]</ref>, a MATLAB implementation of SVM, a complete object oriented environment for machine learning. Being SVM a binary classifier it is necessary to label the datasets for the classification experiments, the beach scenes images are labeled as 1, and the city scene images are labeled as -1.</p><p>In table I there is a list of kernels available in SPIDER, its formula and its parameters.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>VI. EXPERIMENTAL RESULTS</head><p>It is used the "RBF" kernel to execute the experiments with different sigma values. Another parameter used by SPIDER is the 'soft margin parameter', C, which penalizes the training errors. This value is set to 1000 in all the experiments.</p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">The semantic web</title>
		<author>
			<persName><forename type="first">T</forename><surname>Berners-Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Hendler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Lassila</surname></persName>
		</author>
		<ptr target="http://www.scientificamerican.com/article.cfm?id=the-semantic-web" />
	</analytic>
	<monogr>
		<title level="m">Scientific American</title>
				<imprint>
			<date type="published" when="2001-05">2001. May 2001</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">A Framework for Web Science</title>
		<author>
			<persName><surname>Berners-Lee</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Foundations and Trends in Web Science</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="issue">1</biblScope>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Semantic Research for Digital Libraries</title>
		<author>
			<persName><forename type="first">H</forename><surname>Chen</surname></persName>
		</author>
		<ptr target="http://www.dlib.org/dlib/october99/chen/10chen.html" />
	</analytic>
	<monogr>
		<title level="j">D-Lib Magazine</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="issue">10</biblScope>
			<date type="published" when="1999-10">1999. October 1999</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m" type="main">Towards the Semantic Web: Ontology-Driven Knowledge Management</title>
		<editor>Davies J, Fensel D, van Harmelen F</editor>
		<imprint>
			<date type="published" when="2003">2003</date>
			<publisher>John Wiley &amp; Sons, Ltd</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Power of the Few vs. Wisdom of the Crowd: Wikipedia and the Rise of the Bourgeoisie</title>
		<author>
			<persName><forename type="first">A</forename><surname>Kittur</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Chi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">A</forename><surname>Pendleton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Suh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Mytkowicz</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">25th Annual ACM Conference on Human Factors in Computing Systems (CHI 2007)</title>
				<meeting><address><addrLine>New York</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Us vs. Them: Understanding Social Dynamics in Wikipedia with Revert Graph Visualizations</title>
		<author>
			<persName><forename type="first">B</forename><surname>Suh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">H</forename><surname>Chi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">A</forename><surname>Pendleton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kittur</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Visual Analytics Science and Technology</title>
				<meeting><address><addrLine>New York</addrLine></address></meeting>
		<imprint>
			<publisher>IEEE-Press</publisher>
			<date type="published" when="2007">2007</date>
			<biblScope unit="page" from="163" to="170" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Automatic Vandalism Detection in Wikipedia</title>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Anderka</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">30th European Conference on IR Research, ECIR 2008</title>
				<meeting><address><addrLine>Glasgow</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="663" to="668" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Does it matter who contributes: a study on featured articles in the german wikipedia</title>
		<author>
			<persName><forename type="first">K</forename><surname>Stein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Hess</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 18th conference on Hypertext and hypermedia</title>
				<meeting>the 18th conference on Hypertext and hypermedia<address><addrLine>New York</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2007">2007</date>
			<biblScope unit="page" from="171" to="174" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<ptr target="http://www.mkbergman.com/?p=417" />
		<title level="m">99 Wikipedia Sources Aiding the Semantic Web</title>
				<imprint>
			<date>AI3</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">Mining meaning from Wikipedia</title>
		<author>
			<persName><forename type="first">O</forename><surname>Medelyan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Milne</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Legg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">A</forename><surname>Witten</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2008">2008</date>
			<publisher>Hamilton</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Extracting Semantic Relationships between Wikipedia Categories</title>
		<author>
			<persName><forename type="first">S</forename><surname>Chernov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Iofciu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Nejdl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zhou</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 1st Workshop on Semantic Wikis -From Wiki to Semantics</title>
				<meeting>the 1st Workshop on Semantic Wikis -From Wiki to Semantics<address><addrLine>ESWC2006, Budva</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">From Wikipedia to Semantic Relationships: a Semi-automated Annotation Approach</title>
		<author>
			<persName><forename type="first">M</forename><surname>Ruiz-Casado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Alfonseca</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Castells</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 1st Workshop on Semantic Wikis -From Wiki to Semantics</title>
				<meeting>the 1st Workshop on Semantic Wikis -From Wiki to Semantics<address><addrLine>ESWC2006, Budva</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Corpus Exploitation from Wikipedia for Ontology Construction</title>
		<author>
			<persName><forename type="first">G</forename><surname>Cui</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Lu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Chen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Conference on Language Resources and Evaluation</title>
				<meeting><address><addrLine>LREC2008, Morocco</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Automatically Refining the Wikipedia Infobox Ontology</title>
		<author>
			<persName><forename type="first">F</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">S</forename><surname>Weld</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">17 th International World Wide Web Conference</title>
				<meeting><address><addrLine>Beijing</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<title level="m" type="main">Automatic Ontology Extraction for Document Classification</title>
		<author>
			<persName><forename type="first">N</forename><surname>Kozlova</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2006">2006</date>
		</imprint>
		<respStmt>
			<orgName>Saarland University</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Ma. Thesis</note>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<title level="m" type="main">Introducción al estudio de las expresiones fijas</title>
		<author>
			<persName><forename type="first">A</forename><surname>Zuluaga</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1980">1980</date>
			<publisher>Peter Lang</publisher>
			<pubPlace>Frankfurt</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<author>
			<persName><forename type="first">H</forename><surname>Mogorrón</surname></persName>
		</author>
		<title level="m">Los diccionarios electrónicos fraseológicos, perspectivas para la lengua y la traducción</title>
				<imprint>
			<date type="published" when="2004">2004</date>
		</imprint>
		<respStmt>
			<orgName>Universidad de Alicante</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<author>
			<persName><forename type="first">Gómez</forename><surname>De Silva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename></persName>
		</author>
		<title level="m">Diccionario breve de Mexicanismos</title>
				<meeting><address><addrLine>México, FCE.</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2001">2001</date>
		</imprint>
	</monogr>
	<note>1a ed</note>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<author>
			<persName><forename type="first">B</forename><surname>Steel</surname></persName>
		</author>
		<title level="m">Breve Diccionario Ejemplificado de Mexicanismos</title>
				<imprint>
			<date type="published" when="2000">2000</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Lara</surname></persName>
		</author>
		<title level="m">Diccionario del español usual en México</title>
				<imprint>
			<publisher>ISBN</publisher>
			<date type="published" when="2003">2003</date>
			<biblScope unit="page">9789681207045</biblScope>
		</imprint>
	</monogr>
	<note>1a ed</note>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">L</forename><surname>Casado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Agueda</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Agueda</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Corral</surname></persName>
		</author>
		<title level="m">Recopilación de proverbios</title>
				<meeting><address><addrLine>Alcobendas</addrLine></address></meeting>
		<imprint>
			<publisher>ISBN</publisher>
			<date type="published" when="1998">1998</date>
			<biblScope unit="page">9788471436450</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<monogr>
		<title level="m" type="main">Spagnolo-italiano: espressioni idiomatiche e proverbi</title>
		<author>
			<persName><forename type="first">M</forename><surname>Zamora</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1997">1997</date>
			<pubPlace>Milano, EGEA</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><surname>Sevilla</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Cantera</surname></persName>
		</author>
		<title level="m">refranes españoles con su correspondencia catalana, gallega, vasca, francesa e inglesa</title>
				<meeting><address><addrLine>Madrid</addrLine></address></meeting>
		<imprint>
			<publisher>EUNSA</publisher>
			<date type="published" when="1998">1998</date>
			<biblScope unit="volume">877</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><surname>Sevilla</surname></persName>
		</author>
		<title level="m">Divergencias en la traducción de expresiones idiomáticas y refranes (francés-español</title>
				<imprint>
			<date type="published" when="1999">1999</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<monogr>
		<title level="m" type="main">Regular Expression, Pocket referente</title>
		<author>
			<persName><forename type="first">T</forename><surname>Stubblebine</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
	<note>O`really 2nd edition</note>
</biblStruct>

<biblStruct xml:id="b25">
	<monogr>
		<author>
			<persName><forename type="first">P</forename><surname>Procházková</surname></persName>
		</author>
		<title level="m">Fundamentos de la lingüística de corpus, concepción de los corpus y métodos de investigación con corpus</title>
				<imprint>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<monogr>
		<title level="m" type="main">Preliminary Recommendations on Corpus Typology</title>
		<author>
			<persName><forename type="first">J</forename><surname>Sinclair</surname></persName>
		</author>
		<idno>EAG-TCWG-CTYP/P</idno>
		<imprint>
			<date type="published" when="1996">1996</date>
		</imprint>
	</monogr>
	<note type="report_type">EAGLES Document</note>
</biblStruct>

<biblStruct xml:id="b27">
	<monogr>
		<title level="m" type="main">Developing Linguistic Corpora: a Guide to Good Practice Corpus and Text-Basic Principles</title>
		<author>
			<persName><forename type="first">J</forename><surname>Sinclair</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">DecisionGraphs-An Extension of Decision Tres</title>
		<author>
			<persName><forename type="first">J</forename><surname>Oliver</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">TechnicalReport No</title>
		<imprint>
			<biblScope unit="volume">92</biblScope>
			<biblScope unit="page">173</biblScope>
			<date type="published" when="1993">1993</date>
		</imprint>
	</monogr>
	<note>References</note>
</biblStruct>

<biblStruct xml:id="b29">
	<monogr>
		<author>
			<persName><forename type="first">Patrick</forename><surname>Lambe</surname></persName>
		</author>
		<title level="m">Organising Knowledge: Taxonomies, Knowledge and Organisational Effectiveness</title>
				<imprint>
			<publisher>Oxford) Limited. UK</publisher>
			<date type="published" when="2007">2007</date>
			<biblScope unit="page">9781843342274</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<analytic>
		<title level="a" type="main">Dynamic Taxonomies: A Model for Large Information Bases</title>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">M</forename><surname>Sacco</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Knowledge and Data Engineering</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="468" to="479" />
			<date type="published" when="2000-05">May 2000</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b31">
	<analytic>
		<title level="a" type="main">Some Research Results in Dynamic Taxonomy and Faceted Search Systems</title>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">M</forename><surname>Sacco</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">SIGIR&apos;2006 Workshop on Faceted Search</title>
				<meeting><address><addrLine>Seattle, WA, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2006-08">August 2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b32">
	<analytic>
		<title level="a" type="main">MindMap: utilizing multiple taxonomies and visualization to understand a document collection</title>
		<author>
			<persName><forename type="first">S</forename><surname>Spangler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">T</forename><surname>Kreulen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lessler</surname></persName>
		</author>
		<idno>. 1170-1179</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 35th Annual Hawaii International Conference on System Sciences</title>
				<meeting>the 35th Annual Hawaii International Conference on System Sciences</meeting>
		<imprint>
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
	<note>HICSS. Pags</note>
</biblStruct>

<biblStruct xml:id="b33">
	<analytic>
		<title level="a" type="main">Demonstration: Using WordNet to Build Hierarchical Facet Categories</title>
		<author>
			<persName><forename type="first">E</forename><surname>Stoica</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hearst</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">The ACM SIGIR Workshop on Faceted Search</title>
				<imprint>
			<date type="published" when="2006-08">August, 2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b34">
	<monogr>
		<ptr target="http://www.getty.edu/research/conducting_research/standards/cdwa/index.htmll.2009" />
		<title level="m">Categories for the Description of Works of Art (CDWA), editado por Murtha Baca and Patricia Harpring</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b35">
	<monogr>
		<title level="m">El ABC del Arte del siglo XX, Primera edición en español</title>
				<imprint>
			<publisher>Editorial Phaidon Press Limited</publisher>
			<date type="published" when="1999">1999</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b36">
	<monogr>
		<author>
			<persName><surname>Masdearte</surname></persName>
		</author>
		<ptr target="http://www.masdearte.com/item_critica.cfm?id=315.2009" />
		<title level="m">Portal de arte contemporáneo</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b37">
	<analytic>
		<title level="a" type="main">The SemEval-2007 WePS evaluation: Establishing a benchmark for the web people search task</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Tzitzikas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Analyti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Spyratos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Constantopoulos ; Artiles</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gonzalo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Sekine</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the 4th International Workshop on Semantic Evaluations -SemEval 2007, Association for Computational Linguistics</title>
				<meeting>of the 4th International Workshop on Semantic Evaluations -SemEval 2007, Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2007">2007. 2007</date>
			<biblScope unit="volume">62</biblScope>
			<biblScope unit="page" from="64" to="69" />
		</imprint>
	</monogr>
	<note>An Algebra for Specifying Valid Compound Terms in Faceted Taxonomies</note>
</biblStruct>

<biblStruct xml:id="b38">
	<analytic>
		<title level="a" type="main">Weps 2 evaluation campaign: overview of the web people search clustering task</title>
		<author>
			<persName><forename type="first">J</forename><surname>Artiles</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gonzalo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sekine</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the 2nd Web People Search Evaluation Workshop (WePS 2009), 18th WWW Conference</title>
				<meeting>of the 2nd Web People Search Evaluation Workshop (WePS 2009), 18th WWW Conference</meeting>
		<imprint>
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b39">
	<analytic>
		<title level="a" type="main">Principles of hash-based text retrieval</title>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">30th Annual Int. ACM SIGIR Conf</title>
				<editor>
			<persName><forename type="first">Fuhr</forename><surname>Clarke</surname></persName>
		</editor>
		<editor>
			<persName><surname>Kando</surname></persName>
		</editor>
		<editor>
			<persName><surname>Kraaij</surname></persName>
		</editor>
		<editor>
			<persName><surname>Vries</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="2007">2007</date>
			<biblScope unit="page" from="527" to="534" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b40">
	<analytic>
		<title level="a" type="main">Incorporating non-local information into information extraction systems by gibbs sampling</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">R</forename><surname>Finkel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Grenager</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Manning</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the ACL</title>
				<meeting>the ACL</meeting>
		<imprint>
			<date type="published" when="2005">2005</date>
			<biblScope unit="page" from="363" to="370" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b41">
	<analytic>
		<title level="a" type="main">Clustering abstracts of scientific texts using the transition point technique</title>
		<author>
			<persName><forename type="first">D</forename><surname>Pinto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Jiménez-Salazar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the CICLing 2006 Conference</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<meeting>of the CICLing 2006 Conference</meeting>
		<imprint>
			<publisher>Springer-Verlag</publisher>
			<date type="published" when="2006">2006</date>
			<biblScope unit="volume">3878</biblScope>
			<biblScope unit="page" from="536" to="546" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b42">
	<analytic>
		<title level="a" type="main">A comparison of extrinsic clustering evaluation metrics based on formal constraints</title>
		<author>
			<persName><forename type="first">E</forename><surname>Amigó</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gonzalo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Artiles</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Verdejo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Information Retrieval</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="461" to="486" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b43">
	<analytic>
		<author>
			<persName><forename type="first">A</forename><surname>Lopez</surname></persName>
		</author>
		<ptr target="http://ccc.inaoep.mx/cm50-ci10/columna/080721.pdf" />
	</analytic>
	<monogr>
		<title level="m">La busqueda de respuestas, un desafio computacional antiguo y vigente</title>
		<title level="s">La jornada de Oriente</title>
		<imprint>
			<date type="published" when="2008-07">July 2008</date>
			<biblScope unit="page" from="1" to="2" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b44">
	<analytic>
		<title level="a" type="main">The Deployment of a mobile question answering system</title>
		<author>
			<persName><forename type="first">L</forename><surname>Jochen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Search Engine Meeting</title>
				<meeting><address><addrLine>Boston, Massachusetts</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2005-04">April 2005</date>
			<biblScope unit="volume">1</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b45">
	<analytic>
		<title level="a" type="main">English&apos;s Teaching Based On New Technologies</title>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">Zacaras</forename><surname>Flores</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Lozano Torralba</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Cuapa Canto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Vzquez</surname></persName>
		</author>
		<author>
			<persName><surname>Flores</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="s">The International Journal of Technology</title>
		<idno type="ISSN">1832-3669</idno>
		<imprint>
			<date type="published" when="2008">2008</date>
			<publisher>Common Ground Publishing</publisher>
		</imprint>
		<respStmt>
			<orgName>Northeastern University in</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b46">
	<analytic>
		<title level="a" type="main">The computer for the twenty-first century</title>
		<author>
			<persName><forename type="first">M</forename><surname>Weiser</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Scientific American</title>
		<imprint>
			<biblScope unit="page" from="94" to="104" />
			<date type="published" when="1991-09">1991. September</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b47">
	<monogr>
		<author>
			<persName><forename type="first">F</forename><surname>Zacar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Sánchez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Zacar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Méndez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Cuapa</surname></persName>
		</author>
		<title level="m">FINANCIAL MOBILE SYSTEM BASED ON INTELLIGENT AGENTS in the Austrian Computer Society book series</title>
				<meeting><address><addrLine>Austria</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b48">
	<monogr>
		<title level="m" type="main">u-Teacher: Ubiquitous learning approach</title>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">Zacaras</forename><surname>Flores</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Cuapa Canto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Lozano Torralba</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Vzquez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">Zacarias</forename><surname>Flores</surname></persName>
		</author>
		<author>
			<persName><surname>Flores</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2008-06">june 2008</date>
			<biblScope unit="page" from="9" to="20" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b49">
	<analytic>
		<title level="a" type="main">INAOE&apos;s Participation at QA@CLEF</title>
		<author>
			<persName><forename type="first">Alberto</forename><surname>Tellez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Antonio</forename><surname>Juarez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gustavo</forename><surname>Hernandez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Claudia</forename><surname>Denicia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Esau</forename><surname>Villatoro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Manuel</forename><surname>Montes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Luis</forename><surname>Villasenor</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Laboratorio de Tecnologas del Lenguaje, Instituto Nacional de Astrofsica, ptica y Electrnica</title>
				<meeting><address><addrLine>INAOE), Mexico</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b50">
	<analytic>
		<title level="a" type="main">QALL-ME: Question Answering Learning technologies in a multiLingual and multiModal Envinroment</title>
		<author>
			<persName><forename type="first">R</forename><surname>Izquierdo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Ferrndez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ferrndez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Toms</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">L</forename><surname>Vicedo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Martinez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Surez</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Departamento de Lenguajes y Sistemas Informticos</title>
				<imprint/>
		<respStmt>
			<orgName>Universidad de Alicante</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b51">
	<analytic>
		<title level="a" type="main">The Problem with Unstructured Data</title>
		<author>
			<persName><forename type="first">Robert</forename><surname>Blumberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Shaku</forename><surname>Atre</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">DM Review Magazine</title>
		<imprint>
			<date type="published" when="2003-02">February 2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b52">
	<monogr>
		<ptr target="http://bscw.fit.fraunhofer.de/" />
		<title level="m">BSCW (Basic Support for Cooperative Work</title>
				<imprint>
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b53">
	<monogr>
		<ptr target="http://moodle.org" />
		<title level="m">Modular Object-Oriented Dynamic Learning Environment</title>
				<imprint>
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b54">
	<monogr>
		<title level="m" type="main">The emergence of distributed content management and Peer-to-Peer Content Networks</title>
		<imprint>
			<date type="published" when="2001-01">January 2001</date>
			<publisher>Gardner Consulting</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b55">
	<monogr>
		<title/>
		<author>
			<persName><forename type="first">T</forename><surname>Klinberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Manfredi</surname></persName>
		</author>
		<ptr target="http://www.gnutella.com/" />
		<imprint>
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b56">
	<analytic>
		<title level="a" type="main">Handling churn in a DHT</title>
		<author>
			<persName><forename type="first">S</forename><surname>Rhea</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Geels</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Roscoe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kubiatowicz</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the use USENIX annual technical conference</title>
				<meeting>of the use USENIX annual technical conference</meeting>
		<imprint>
			<date type="published" when="2004-06">June 2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b57">
	<monogr>
		<title/>
		<author>
			<persName><forename type="first">Jxta</forename><surname>Juxtapose</surname></persName>
		</author>
		<ptr target="https://jxta.dev.java.net/2009" />
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b58">
	<analytic>
		<title level="a" type="main">Recomendadores en un Sistema Multiagente Adaptativo para el Comercio Electrónico</title>
		<author>
			<persName><forename type="first">Ana</forename><surname>Gil</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Zahia</forename><surname>Guessoum</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Francisco</forename><surname>García</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Taller en Sistemas Hipermedia Colaborativos y Adaptativos dentro de las JISBD&apos;2002</title>
				<meeting><address><addrLine>El Escorial</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2002">22 de Noviembre del 2002</date>
			<biblScope unit="volume">18</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b59">
	<analytic>
		<title level="a" type="main">A Recommender Agent Design For Virtual Worlds</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">C</forename><surname>Ramirez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Vilariño</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>López</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IASK International Conference -E-Activity and Leading Technologies</title>
				<meeting>the IASK International Conference -E-Activity and Leading Technologies</meeting>
		<imprint>
			<date type="published" when="2009">2009</date>
			<biblScope unit="page" from="142" to="146" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b60">
	<monogr>
		<ptr target="http://opensimulator.orgSomeConsiderationsfortheSemanticWeb" />
		<title level="m">OpenSimulator Web</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b61">
	<monogr>
		<title level="m" type="main">Draft bioinformatics. Hybrid techniques called K-NM-PSO-based K-means, Nelder-Mead Simplex search and optimization of exchange of particles</title>
		<author>
			<persName><forename type="first">Yi-Tung</forename><surname>Kao</surname></persName>
		</author>
		<author>
			<persName><surname>Zahara</surname></persName>
		</author>
		<author>
			<persName><surname>Erwie</surname></persName>
		</author>
		<author>
			<persName><surname>Kao</surname></persName>
		</author>
		<author>
			<persName><surname>I-Wei</surname></persName>
		</author>
		<imprint/>
	</monogr>
	<note>A hibridized approach to data clustering</note>
</biblStruct>

<biblStruct xml:id="b62">
	<monogr>
		<title level="m" type="main">Enhancing K-Means Algorithm with Initial Cluster Centers Derived from Data Partitioning along the Data Axis with the Highest Variance</title>
		<author>
			<persName><forename type="first">S</forename><surname>Deelers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>And</surname></persName>
		</author>
		<author>
			<persName><surname>Auwatanamongkol</surname></persName>
		</author>
		<imprint/>
	</monogr>
	<note>Title explicit</note>
</biblStruct>

<biblStruct xml:id="b63">
	<monogr>
		<title level="m" type="main">A method for initialising the K-means clustering algorithm using kd-trees&quot; A kd-tree used to calculate an estimate of the density of data and to select the number of clusters</title>
		<author>
			<persName><forename type="first">Stephen</forename><forename type="middle">J</forename><surname>Redmond</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Conor</forename><surname>Heneghan</surname></persName>
		</author>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b64">
	<analytic>
		<title level="a" type="main">Analysis of Global k-means, an Incremental Heuristic for Minimum Sum of Squares Clustering</title>
		<author>
			<persName><forename type="first">P</forename><surname>Hansen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Nagai</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Commentary on work</title>
		<imprint>
			<biblScope unit="volume">22</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b65">
	<monogr>
		<title level="m" type="main">to select the reference number of clusters</title>
		<author>
			<persName><forename type="first">D</forename><surname>Pham</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">S</forename><surname>Dimov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">D</forename><surname>Nguyen</surname></persName>
		</author>
		<imprint/>
	</monogr>
	<note>Selection of K in K-means clustering</note>
</biblStruct>

<biblStruct xml:id="b66">
	<analytic>
		<title level="a" type="main">of initial partitions and accelerates the calculations of k-means</title>
		<author>
			<persName><forename type="first">A</forename><surname>Likas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Vlassis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">J</forename><surname>Verbeek</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Algorithm that consists of a series of k-means clusterings with varying number of clusters</title>
				<imprint/>
	</monogr>
	<note>The Global K-means Clustering Algorithm</note>
</biblStruct>

<biblStruct xml:id="b67">
	<analytic>
		<title level="a" type="main">An empirical comparision of four initialization methods for the k-means algorithm</title>
		<author>
			<persName><forename type="first">J</forename><surname>Peña</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lozano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Larrañaga</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Random</title>
		<imprint/>
	</monogr>
	<note>Compare initialization methods for k-means. 20</note>
</biblStruct>

<biblStruct xml:id="b68">
	<monogr>
		<title level="m" type="main">Use k-means M times for M random subsets of the original data</title>
		<author>
			<persName><forename type="first">P</forename><surname>Bradley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Fayyad</surname></persName>
		</author>
		<imprint/>
	</monogr>
	<note>Refining initial points for k-means clustering</note>
</biblStruct>

<biblStruct xml:id="b69">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Kaufman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">L</forename><surname>Rouseeuw</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Kaufman</surname></persName>
		</author>
		<author>
			<persName><surname>Rouseeuw</surname></persName>
		</author>
		<title level="m">Finding Groups in Data: An Introduction to Cluster analysis</title>
				<imprint/>
	</monogr>
	<note>Text. Def. K-means</note>
</biblStruct>

<biblStruct xml:id="b70">
	<analytic>
		<title level="a" type="main">A clustering technique for summarizing multivariate data</title>
		<author>
			<persName><forename type="first">G</forename><surname>Ball</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Hall</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ISODATA). Perform dynamic estimation of K</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b71">
	<analytic>
		<title level="a" type="main">to perform global search and rapid convergence</title>
		<author>
			<persName><forename type="first">K</forename><surname>Krishna</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Murty</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Hybrid scheme based on Genetic Algorithm -Simulated annealing with new operators</title>
				<imprint/>
	</monogr>
	<note>Genetic K-means algorithm</note>
</biblStruct>

<biblStruct xml:id="b72">
	<analytic>
		<title level="a" type="main">Some Methods for Classification and Analysis of Multivariate Observations</title>
		<author>
			<persName><forename type="first">J</forename><surname>Macqueen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Definition, Analysis and Applications of kmeans</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b73">
	<monogr>
		<title level="m" type="main">(n, d, k) which involves the sample size, the number of dimensions and the number of partitions. There are</title>
		<imprint/>
	</monogr>
	<note>2.3 The efficiency of the algorithm. According to the work of [42] the complexity of the k-means algorithm is O. several works that have focused on different aspects of the algorithm, in order to reduce computational load. Authors Títle and Commentary</note>
</biblStruct>

<biblStruct xml:id="b74">
	<analytic>
		<title level="a" type="main">New Efficient Strategy to Accelerate k-Means Clustering Algorithm</title>
		<author>
			<persName><forename type="first">`</forename><surname>Moh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Belal</forename><surname>Al-Zoubi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Amjad</forename><surname>Hudaib</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ammar</forename><surname>Huneiti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Bassam</forename><surname>Hammo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Strategy to accelerate k-means algorithm, which avoids many calculations of distance, through a strategy based on an improvement to the partial distance algorithm</title>
				<imprint/>
	</monogr>
	<note>PD</note>
</biblStruct>

<biblStruct xml:id="b75">
	<monogr>
		<title level="m" type="main">An Efficient k`-means Clustering Algorithm</title>
		<author>
			<persName><forename type="first">Krista</forename><surname>Zalik</surname></persName>
		</author>
		<author>
			<persName><surname>Rizman</surname></persName>
		</author>
		<imprint/>
	</monogr>
	<note>Based on the algorithm Rival, it penalizes competitive Learninig (RPCL). It does not require pre-allocation of the number of clusters. Two-step process. Pre processes and uses the prior information to minimize the cost function</note>
</biblStruct>

<biblStruct xml:id="b76">
	<monogr>
		<title level="m" type="main">Eliminates the need to specify a priori the number of clusters. Combines genetic algorithms and logarithmic regression Màxima expectation</title>
		<author>
			<persName><forename type="middle">D</forename><surname>Cao</surname></persName>
		</author>
		<author>
			<persName><surname>Nguyen &amp; Cios</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Krzysztof</surname></persName>
		</author>
		<imprint/>
	</monogr>
	<note>GAKREM: A novel hybrid clustering algorithm</note>
</biblStruct>

<biblStruct xml:id="b77">
	<monogr>
		<title level="m" type="main">A Fast k-means implementation using coresets</title>
		<author>
			<persName><forename type="first">G</forename><surname>Frahling</surname></persName>
		</author>
		<author>
			<persName><surname>Ch</surname></persName>
		</author>
		<author>
			<persName><surname>Sohler</surname></persName>
		</author>
		<imprint/>
	</monogr>
	<note>Implemented version of Lloyd&apos;s k-means [23], using a weighted set of points that approximate the original set</note>
</biblStruct>

<biblStruct xml:id="b78">
	<monogr>
		<title level="m" type="main">Improvement of the algorithm by introducing a variable to the function of cost</title>
		<author>
			<persName><forename type="first">Taoying</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yan</forename><surname>Chen</surname></persName>
		</author>
		<imprint/>
	</monogr>
	<note>An improved k-means algorithm for clustering using entropy weighting measures</note>
</biblStruct>

<biblStruct xml:id="b79">
	<analytic>
		<title level="a" type="main">K-means clustering of proportional data using L1 distance</title>
		<author>
			<persName><forename type="first">H</forename><surname>Kashima</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B;</forename><surname>Ray</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Singh</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proportionate restrictions incorporated in the calculation of centroids</title>
				<imprint/>
	</monogr>
	<note>K-means based on distance L1</note>
</biblStruct>

<biblStruct xml:id="b80">
	<analytic>
		<title level="a" type="main">Developing a feature weight self-adjustment mechanism for a K-means clustering algorithm</title>
		<author>
			<persName><forename type="first">Chieh-</forename><surname>Tsai</surname></persName>
		</author>
		<author>
			<persName><surname>Yuan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Chuang-</forename><surname>Chiu</surname></persName>
		</author>
		<author>
			<persName><surname>Cheng</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Improvement of the quality of k-means clustering via FWSA mechanism called</title>
				<imprint/>
	</monogr>
	<note>Is modeled as an optimization problem</note>
</biblStruct>

<biblStruct xml:id="b81">
	<analytic>
		<title level="a" type="main">Improvement the Efficiency and Efficacy of the K-means Clustering Algorithm through a New Convergence Condition</title>
		<author>
			<persName><forename type="first">Joaquín</forename><surname>Pérez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename></persName>
		</author>
		<author>
			<persName><forename type="first">Rodolfo</forename><surname>Pazos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename></persName>
		</author>
		<author>
			<persName><forename type="first">Laura</forename><surname>Cruz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename></persName>
		</author>
		<author>
			<persName><forename type="first">Gerardo</forename><surname>Reyes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Rosy Basave</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Héctor Fraire</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Improvement to the k-means algorithm by new convergence conditions</title>
				<imprint/>
	</monogr>
	<note>of kmeans</note>
</biblStruct>

<biblStruct xml:id="b82">
	<monogr>
		<title level="m" type="main">Improvement of the K-means algorithm using a new approach of convergence and its application to databases cancer population</title>
		<author>
			<persName><forename type="first">J</forename><surname>Pérez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">F</forename><surname>Henriques</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Pazos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Cruz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Reyes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Salinas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mexicano</surname></persName>
		</author>
		<imprint/>
	</monogr>
	<note>Title explicit</note>
</biblStruct>

<biblStruct xml:id="b83">
	<analytic>
		<title level="a" type="main">Unique distance measure approach for K-means (UDMA-Km) clustering algorithm</title>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">K D</forename><surname>Pun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">S</forename><surname>Ali</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Sets distance measure based on statistical data</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b84">
	<monogr>
		<title level="m" type="main">Define new objective function and minimize it with genetic algorithm</title>
		<author>
			<persName><forename type="first">Zejin</forename><surname>Ding</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jian</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Self-Yang-Qing</forename><surname>Zhang</surname></persName>
		</author>
		<imprint/>
	</monogr>
	<note>A New improved K-Means Algorithm with Penalizaed Term</note>
</biblStruct>

<biblStruct xml:id="b85">
	<analytic>
		<title level="a" type="main">An Efficient K-means Clustering Algorithm: Analysis and Implementation</title>
		<author>
			<persName><forename type="first">T</forename><surname>Kanungo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">M</forename><surname>Mount</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">S</forename><surname>Netanyahu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">D</forename><surname>Piatko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Silverman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">Y</forename><surname>Wu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">presents an implementation of the version of Lloyd</title>
				<imprint/>
	</monogr>
	<note>&apos;s k-means [23] called &quot;filtering algorithm&quot; based on a kd-tree</note>
</biblStruct>

<biblStruct xml:id="b86">
	<monogr>
		<title/>
		<author>
			<persName><forename type="first">K-</forename></persName>
		</author>
		<imprint/>
	</monogr>
	<note>means is sensitive to outliers and noise. According to [42] even if an object is quite far away from the cluster centroid, it is still forced into a cluster and, thus, it distorts the cluster shapes. Here are works that focus on shortcoming: References</note>
</biblStruct>

<biblStruct xml:id="b87">
	<analytic>
		<title level="a" type="main">Geodesic K-means clustering</title>
		<author>
			<persName><forename type="first">N</forename><surname>Asgharbeygi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Maleki</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ICPR 2008, 19 th International Conference on</title>
				<imprint>
			<date type="published" when="2008-12">2008. Dec. 2008</date>
		</imprint>
	</monogr>
	<note>Pattern Recognition</note>
</biblStruct>

<biblStruct xml:id="b88">
	<analytic>
		<title level="a" type="main">A New Evolutionary Algorithm for Cluster Analysis</title>
		<author>
			<persName><forename type="first">B</forename><surname>Bahmani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Firouzi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Niknam</surname></persName>
		</author>
		<author>
			<persName><surname>Nayeripour</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Proceedings of world Academy of Science, Engineering and Technology</title>
		<imprint>
			<biblScope unit="volume">36</biblScope>
			<date type="published" when="2008-12">Dec.2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b89">
	<analytic>
		<title level="a" type="main">A clustering technique for summarizing multivariate data</title>
		<author>
			<persName><forename type="first">G</forename><surname>Ball</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Hall</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Behav Sci</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="page" from="153" to="155" />
			<date type="published" when="1967">1967</date>
		</imprint>
	</monogr>
	<note>ISODATA)</note>
</biblStruct>

<biblStruct xml:id="b90">
	<analytic>
		<title level="a" type="main">New Efficient Strategy to Accelerate k-Means Clustering Algorithm</title>
		<author>
			<persName><forename type="first">Belal</forename><surname>Al-Zoubi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Amjad</forename><surname>Al-Zoubi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ammar</forename><surname>Hudaib</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Bassam</forename><surname>Huneiti</surname></persName>
		</author>
		<author>
			<persName><surname>Hammo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">American Journal of Applied Sciences</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="issue">9</biblScope>
			<biblScope unit="page" from="1247" to="1250" />
			<date type="published" when="2008">2008</date>
			<publisher>Science Publications</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b91">
	<analytic>
		<title level="a" type="main">Refining initial points for k-means clustering</title>
		<author>
			<persName><forename type="first">P</forename><surname>Bradley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Fayyad</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. 15 th Int. Conf. Machine Learning</title>
				<meeting>15 th Int. Conf. Machine Learning</meeting>
		<imprint>
			<date type="published" when="1998">1998</date>
			<biblScope unit="page" from="91" to="99" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b92">
	<analytic>
		<title level="a" type="main">Enhancing K-Means Algorithm with Initial Cluster Centers Derived from Data Partitioning along the Data Axis with the Highest Variance</title>
		<author>
			<persName><forename type="first">S</forename><surname>Deelers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>And</surname></persName>
		</author>
		<author>
			<persName><surname>Auwatanamongkol</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Proceedings of world Academy of Science, Engineering and Technology</title>
		<imprint>
			<biblScope unit="volume">26</biblScope>
			<date type="published" when="2007-12">Dec. 2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b93">
	<analytic>
		<title level="a" type="main">A New improved K-Means Algorithm with Penalizaed Term</title>
		<author>
			<persName><forename type="first">Ding</forename><surname>Zejin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jian</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yang-Qing</forename><surname>Zhang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE International Conference on</title>
				<imprint>
			<date type="published" when="2007-11">2007. Nov. 2007</date>
		</imprint>
	</monogr>
	<note>Granular Computing</note>
</biblStruct>

<biblStruct xml:id="b94">
	<monogr>
		<title level="m" type="main">Pattern Classification and Scene Analysis</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">O</forename><surname>Duda</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">E</forename><surname>Hart</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1973">1973</date>
			<publisher>John Wiley &amp;Sons</publisher>
			<pubPlace>New York, NY</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b95">
	<analytic>
		<title level="a" type="main">A fast and robust general purpose clustering algorithm</title>
		<author>
			<persName><forename type="first">V</forename><surname>Estivill-Castro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Yang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. 6 th Pacific Rim Int. Conf. Artificial Intelligence (PRICAI`00</title>
				<editor>
			<persName><forename type="first">R</forename><surname>Mizoguchi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Slaney</surname></persName>
		</editor>
		<meeting>6 th Pacific Rim Int. Conf. Artificial Intelligence (PRICAI`00<address><addrLine>Melbourne, Australia</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2000">2000</date>
			<biblScope unit="page" from="208" to="218" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b96">
	<monogr>
		<author>
			<persName><forename type="first">U</forename><forename type="middle">M</forename><surname>Fayyad</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Piatetsky-Shanpiro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Smyth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Uthurusamy</surname></persName>
		</author>
		<title level="m">Advances in Knowledge Discovery and Data Mining</title>
				<imprint>
			<publisher>AAAI/MIT Press</publisher>
			<date type="published" when="1996">1996</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b97">
	<analytic>
		<title level="a" type="main">Knowledge Acquisition via Incremental Conceptual Clustering</title>
		<author>
			<persName><forename type="first">D</forename><surname>Fissher</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Machine Learning</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="139" to="172" />
			<date type="published" when="1987">1987</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b98">
	<analytic>
		<title level="a" type="main">Cluster analysis of multivariate data: Efficiency vs. Interpretability of classification</title>
		<author>
			<persName><forename type="first">E</forename><surname>Forgy</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Biometrics</title>
		<imprint>
			<biblScope unit="volume">21</biblScope>
			<biblScope unit="page" from="768" to="780" />
			<date type="published" when="1965">1965</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b99">
	<analytic>
		<title level="a" type="main">A Fast k-means implementation using coresets</title>
		<author>
			<persName><forename type="first">G</forename><surname>Frahling</surname></persName>
		</author>
		<author>
			<persName><surname>Ch</surname></persName>
		</author>
		<author>
			<persName><surname>Sohler</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Computational Geometry &amp; Applications</title>
		<imprint>
			<biblScope unit="volume">18</biblScope>
			<biblScope unit="issue">6</biblScope>
			<biblScope unit="page" from="P605" to="625" />
			<date type="published" when="2008-12">Dec. 2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b100">
	<analytic>
		<title level="a" type="main">K-means clustering algorithm for categorical attributes</title>
		<author>
			<persName><forename type="first">S</forename><surname>Gupata</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Rao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Bhatnagar</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. 1 st Int. Conf. Data Werehousing and Knowledge Discovery (DaWak`99)</title>
				<meeting>1 st Int. Conf. Data Werehousing and Knowledge Discovery (DaWak`99)<address><addrLine>Florence, Italy</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1999">1999</date>
			<biblScope unit="page" from="203" to="208" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b101">
	<analytic>
		<title level="a" type="main">Analysis of Global k-means, an Incremental Heuristic for Minimum Sum of Squares Clustering</title>
		<author>
			<persName><forename type="first">P</forename><surname>Hansen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Nagai</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal Classification</title>
		<imprint>
			<biblScope unit="volume">22</biblScope>
			<biblScope unit="page" from="287" to="310" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b102">
	<analytic>
		<title level="a" type="main">Extensions to the k-means algorithm for clustering large data sets with categorical values</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Huang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Data Mining Knowl. Discov</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="283" to="304" />
			<date type="published" when="1998">1998</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b103">
	<analytic>
		<title level="a" type="main">An Efficient K-means Clustering Algorithm: Analysis and Implementation</title>
		<author>
			<persName><forename type="first">T</forename><surname>Kanungo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">M</forename><surname>Mount</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">S</forename><surname>Netanyahu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">D</forename><surname>Piatko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Silverman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">Y</forename><surname>Wu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Pattern Analysis and Machine Intelligence</title>
		<imprint>
			<biblScope unit="volume">24</biblScope>
			<biblScope unit="issue">7</biblScope>
			<biblScope unit="page" from="881" to="892" />
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
	<note>Pattern Analysis and Machine Intelligence</note>
</biblStruct>

<biblStruct xml:id="b104">
	<analytic>
		<title level="a" type="main">K-means clustering of proportional data using L1 distance</title>
		<author>
			<persName><forename type="first">H</forename><surname>Kashima</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B;</forename><surname>Ray</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Singh</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference On</title>
				<imprint>
			<date type="published" when="2008-12">2008. Dec. 2008</date>
		</imprint>
	</monogr>
	<note>Pattern Recognition</note>
</biblStruct>

<biblStruct xml:id="b105">
	<analytic>
		<title level="a" type="main">A hibridized approach to data clustering</title>
		<author>
			<persName><forename type="first">Yi-Tung</forename><surname>Kao</surname></persName>
		</author>
		<author>
			<persName><surname>Zahara</surname></persName>
		</author>
		<author>
			<persName><surname>Erwie</surname></persName>
		</author>
		<author>
			<persName><surname>Kao</surname></persName>
		</author>
		<author>
			<persName><surname>I-Wei</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Expert Systems with Applications</title>
		<imprint>
			<biblScope unit="volume">34</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="1754" to="1762" />
			<date type="published" when="2008-04">Apr.2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b106">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Kaufman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rouseeuw</surname></persName>
		</author>
		<title level="m">Finding Groups in Data: An Introduction to Cluster analysis</title>
				<imprint>
			<publisher>Wiley</publisher>
			<date type="published" when="1990">1990</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b107">
	<analytic>
		<title level="a" type="main">Genetic K-means algorithm</title>
		<author>
			<persName><forename type="first">K</forename><surname>Krishna</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Murty</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Trans. Syst., Man, Cybern. B., Cybern</title>
		<imprint>
			<biblScope unit="volume">29</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="433" to="439" />
			<date type="published" when="1999-06">Jun. 1999</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b108">
	<analytic>
		<title level="a" type="main">The Global K-means Clustering Algorithm</title>
		<author>
			<persName><forename type="first">A</forename><surname>Likas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Vlassis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">J</forename><surname>Verbeek</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">The Journal of the Pattern Recognition Society</title>
		<imprint>
			<biblScope unit="volume">36</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="451" to="461" />
			<date type="published" when="2003">2003</date>
		</imprint>
	</monogr>
	<note>Pattern Recognition</note>
</biblStruct>

<biblStruct xml:id="b109">
	<analytic>
		<title level="a" type="main">Least squares quantization in PCM</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">P</forename><surname>Lloyd</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">portions presented at the Institute of Mathematical statistics Meeting Atlantic</title>
				<meeting><address><addrLine>City, NJ; Mach</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1957-09">sep. 1957. 1982</date>
			<biblScope unit="volume">28</biblScope>
			<biblScope unit="page" from="129" to="137" />
		</imprint>
		<respStmt>
			<orgName>Unpublished Bell Lab. Tech. Note</orgName>
		</respStmt>
	</monogr>
	<note>IEEE Trans. inform, Theory (Special Issue on Quantization)</note>
</biblStruct>

<biblStruct xml:id="b110">
	<analytic>
		<title level="a" type="main">Some Methods for Classification and Analysis of Multivariate Observations</title>
		<author>
			<persName><forename type="first">J</forename><surname>Macqueen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings Fifth Berkeley Symposium Mathematics Statistics and Probability</title>
				<meeting>Fifth Berkeley Symposium Mathematics Statistics and Probability</meeting>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b111">
	<monogr>
		<title/>
		<author>
			<persName><surname>Matworks</surname></persName>
		</author>
		<ptr target="http://www.matworks.com" />
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b112">
	<monogr>
		<title level="m" type="main">Data Mining: Concepts, Models, Methods, and Algorithms</title>
		<author>
			<persName><forename type="first">K</forename><surname>Mehmed</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2003">2003</date>
			<publisher>John Wiley &amp; Sons</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b113">
	<analytic>
		<title level="a" type="main">GAKREM: A novel hybrid clustering algorithm</title>
		<author>
			<persName><forename type="first">Cao</forename><forename type="middle">D</forename><surname>Nguyen</surname></persName>
		</author>
		<author>
			<persName><surname>Cios</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Krzysztof</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Information Sciences</title>
		<imprint>
			<biblScope unit="volume">178</biblScope>
			<biblScope unit="issue">22</biblScope>
			<biblScope unit="page" from="4205" to="4227" />
			<date type="published" when="2008-11">Nov. 2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b114">
	<analytic>
		<title level="a" type="main">An empirical comparision of four initialization methods for the k-means algorithm</title>
		<author>
			<persName><forename type="first">J</forename><surname>Peña</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Lozano</surname></persName>
		</author>
		<author>
			<persName><surname>Larrañaga</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Pattern Recognit Lett</title>
		<imprint>
			<biblScope unit="volume">20</biblScope>
			<biblScope unit="page" from="1027" to="1040" />
			<date type="published" when="1999">1999</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b115">
	<analytic>
		<title level="a" type="main">Improvement the Efficiency and Efficacy of the K-means Clustering Algorithm through a New Convergence Condition</title>
		<author>
			<persName><forename type="first">J</forename><surname>Pérez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Rodolfo</forename><surname>Pazos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename></persName>
		</author>
		<author>
			<persName><forename type="first">Laura</forename><surname>Cruz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename></persName>
		</author>
		<author>
			<persName><forename type="first">Gerardo</forename><surname>Reyes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Rosy Basave</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Héctor Fraire</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Computational Science and Its Applications -ICCSA 2007 -International Conference Proceedings</title>
				<imprint>
			<publisher>Springer Verlag</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b116">
	<analytic>
		<title level="a" type="main">Mejora al Algoritmo de K-means mediante un Nuevo criterio de convergencia y su aplicación a bases de datos poblacionales de cáncer. 2do Taller Latino Iberoamericano de Investigación de Operaciones</title>
		<author>
			<persName><forename type="first">J</forename><surname>Pérez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">F</forename><surname>Henriques</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Pazos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Cruz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Reyes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Salinas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mexicano</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">La IO aplicada a la solución de problemas regionales</title>
				<meeting><address><addrLine>México</addrLine></address></meeting>
		<imprint/>
	</monogr>
	<note>In Spanish</note>
</biblStruct>

<biblStruct xml:id="b117">
	<analytic>
		<title level="a" type="main">Selection of K in K-means clustering</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">T</forename><surname>Pham</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">S</forename><surname>Dimov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">D</forename><surname>Nguyen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Proceedings of the Institution of Mechanical Engineers -Part C -Journal of Mechanical Engineering Science</title>
		<imprint>
			<biblScope unit="volume">219</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="103" to="109" />
			<date type="published" when="2005-01">Jan 2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b118">
	<analytic>
		<title level="a" type="main">Analysis of Range Queries on Real Region Datasets Stored Using an R-Tree</title>
		<author>
			<persName><forename type="first">Guido</forename><surname>Proietti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Christos</forename><surname>Faloutsos</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Knowledge and Data Engieneering</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="issue">5</biblScope>
			<date type="published" when="2000-10">Sep./Oct. 2000</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b119">
	<analytic>
		<title level="a" type="main">Unique distance measure approach for K-means (UDMA-Km) clustering algorithm</title>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">K D</forename><surname>Pun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">S</forename><surname>Ali</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">TENCON 2007 -2007 IEEE Region 10 Conference</title>
				<imprint>
			<date type="published" when="2007">Oct. 30 2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b120">
	<analytic>
		<title level="a" type="main">A method for initialising the K-means clustering algorithm using kd-trees</title>
		<author>
			<persName><forename type="first">Stephen</forename><forename type="middle">J</forename><surname>Redmond</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Conor</forename><surname>Heneghan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Pattern Recognition Letters</title>
		<imprint>
			<biblScope unit="volume">28</biblScope>
			<biblScope unit="issue">8</biblScope>
			<date type="published" when="2007-06">Jun. 2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b121">
	<analytic>
		<title level="a" type="main">An improved k-means algorithm for clustering using entropy weighting measures</title>
		<author>
			<persName><forename type="first">Taoying</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yan</forename><surname>Chen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Intelligent Control and Automation</title>
				<meeting><address><addrLine>WCICA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2008-06">2008. 2008. June 2008</date>
		</imprint>
	</monogr>
	<note>7 th World Congress on</note>
</biblStruct>

<biblStruct xml:id="b122">
	<analytic>
		<title level="a" type="main">Developing a feature weight selfadjustment mechanism for a K-means clustering algorithm</title>
		<author>
			<persName><forename type="first">Chieh-</forename><surname>Tsai</surname></persName>
		</author>
		<author>
			<persName><surname>Yuan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Chuang-</forename><surname>Chiu</surname></persName>
		</author>
		<author>
			<persName><surname>Cheng</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computational Statistics &amp; Data Analysis</title>
		<imprint>
			<biblScope unit="volume">52</biblScope>
			<biblScope unit="page">10</biblScope>
			<date type="published" when="2008-06">Jun. 2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b123">
	<monogr>
		<title level="m" type="main">UCI Machine Learning Repository</title>
		<author>
			<persName><forename type="first">A</forename><surname>Uci. Asuncion</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">J</forename><surname>Newman</surname></persName>
		</author>
		<ptr target="http://www.ics.uci.edu/~mlearn/MLRepository.html" />
		<imprint>
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b124">
	<analytic>
		<title level="a" type="main">Genetic Algorithm for text clustering using ontology and evaluating the vality of various semantic simility measures</title>
		<author>
			<persName><forename type="first">Song</forename><surname>Wei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Li</forename><forename type="middle">Cheng</forename><surname>Hua</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Soon</forename><surname>Park</surname></persName>
		</author>
		<author>
			<persName><surname>Cheo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Expert Systems with Applications</title>
		<imprint>
			<biblScope unit="volume">36</biblScope>
			<biblScope unit="issue">5</biblScope>
			<date type="published" when="2009-07">Jul. 2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b125">
	<analytic>
		<title level="a" type="main">Local vs global interactions in clustering algorithms: Advances over K-means</title>
		<author>
			<persName><forename type="first">Barbakh</forename><surname>Wesan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Colin</forename><surname>And</surname></persName>
		</author>
		<author>
			<persName><surname>Fyfe</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of knowledge-based and Intelilligent Engineering Systems</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="page" from="83" to="99" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b126">
	<monogr>
		<title level="m" type="main">Data Mining: Practical Machine Learning Tools and Techniques with Java Implementations</title>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">H</forename><surname>Witten</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Frank</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1999">1999</date>
			<publisher>Morgan Kaufmann Publishers</publisher>
			<pubPlace>San Diego, CA</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b127">
	<analytic>
		<title level="a" type="main">Top 10 algorthms in data mining</title>
		<author>
			<persName><forename type="first">Xindong</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">,</forename></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Kumar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ross</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ghosh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Motoda</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">J</forename><surname>Mclachlan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">S</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Steinbach</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">J</forename><surname>Hand</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Steinberg</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Knowl Inf Syst</title>
		<imprint>
			<date type="published" when="2008">2008</date>
			<publisher>Springer</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b128">
	<analytic>
		<title level="a" type="main">Survey of Clustering Algorithms</title>
		<author>
			<persName><forename type="first">Rui</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Donald</forename><surname>Wunsch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">I</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Neural Networks</title>
		<imprint>
			<biblScope unit="volume">16</biblScope>
			<biblScope unit="issue">3</biblScope>
			<date type="published" when="2005-05">May 2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b129">
	<analytic>
		<title level="a" type="main">An Efficient k`-means Clustering Algorithm</title>
		<author>
			<persName><forename type="first">Krista</forename><surname>Zalik</surname></persName>
		</author>
		<author>
			<persName><surname>Rizman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Pattern Reconition Letters</title>
		<imprint>
			<biblScope unit="volume">29</biblScope>
			<biblScope unit="issue">9</biblScope>
			<biblScope unit="page" from="1385" to="1391" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
	<note>Elsevier</note>
</biblStruct>

<biblStruct xml:id="b130">
	<analytic>
		<title level="a" type="main">On the Lower Bound of Local Optimums in K-means Algorithm</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>B. Tian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">K H</forename><surname>Tung</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ICDM`06 Sixth International Conference on Data Mining</title>
				<imprint>
			<date type="published" when="2006-12">Dec. 2006</date>
		</imprint>
	</monogr>
	<note>Data Mining 2006</note>
</biblStruct>

<biblStruct xml:id="b131">
	<analytic>
		<title level="a" type="main">K-means Clustering Algorithm with improved initial Center</title>
		<author>
			<persName><forename type="first">Chen; Xia</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><surname>Shixiong</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Knowledge Discovery and Data Mining</title>
				<imprint>
			<date type="published" when="2009-01">2009. Jan., 2009</date>
			<biblScope unit="page" from="23" to="25" />
		</imprint>
	</monogr>
	<note>Second International Workshop on</note>
</biblStruct>

<biblStruct xml:id="b132">
	<analytic>
		<title level="a" type="main">Filtering for texture classification: a comparative study</title>
		<author>
			<persName><forename type="first">T</forename><surname>Randen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">H</forename><surname>Husoy</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Trans. on Pattern Analysis and Machine Intelligence</title>
		<imprint>
			<biblScope unit="volume">21</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="291" to="310" />
			<date type="published" when="1999-04">Apr 1999</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b133">
	<monogr>
		<title level="m" type="main">Segmentation, classification and modelization of textures by means of multiresolution decomposition techniques</title>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">Lumbreras</forename><surname>Ruiz</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2001">2001</date>
			<pubPlace>Barcelona, España</pubPlace>
		</imprint>
		<respStmt>
			<orgName>Dept. Informática and Computer Vision Center, Universitat Autònoma de Barcelona</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Ph.D. dissertation</note>
</biblStruct>

<biblStruct xml:id="b134">
	<analytic>
		<title level="a" type="main">Support vector machines for histogram-based image classification</title>
		<author>
			<persName><forename type="first">O</forename><surname>Chapelle</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Haffner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">N</forename><surname>Vapnik</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Trans. On Neural Networks</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="issue">5</biblScope>
			<biblScope unit="page" from="1055" to="1064" />
			<date type="published" when="1999-09">Sep 1999</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b135">
	<analytic>
		<title level="a" type="main">Support vector machines for texture classification</title>
		<author>
			<persName><forename type="first">In</forename><surname>Kwang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Keechul</forename><surname>Kim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Se</forename><surname>Jung</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hyun</forename><surname>Park</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hang</forename><forename type="middle">Joon</forename><surname>Kim</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Trans. On Pattern Analysis and Machine Intelligence</title>
		<imprint>
			<biblScope unit="volume">24</biblScope>
			<biblScope unit="issue">11</biblScope>
			<biblScope unit="page" from="1542" to="1550" />
			<date type="published" when="2002-11">Nov 2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b136">
	<analytic>
		<title level="a" type="main">Learning in region-based image retrieval with generalized support vector machines</title>
		<author>
			<persName><forename type="first">I</forename><surname>Gondra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">R</forename><surname>Heisterkamp</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the Computer Vision and Pattern Recognition</title>
				<meeting>of the Computer Vision and Pattern Recognition</meeting>
		<imprint>
			<date type="published" when="2004">2004</date>
			<biblScope unit="page" from="149" to="154" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b137">
	<analytic>
		<title level="a" type="main">Texture classification using the support vector machines</title>
		<author>
			<persName><forename type="first">Shutao</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">T</forename><surname>Kwok</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hailong</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yaonan</forename><surname>Wang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Pattern Recognition</title>
		<imprint>
			<biblScope unit="volume">36</biblScope>
			<biblScope unit="issue">12</biblScope>
			<biblScope unit="page" from="2883" to="2893" />
			<date type="published" when="2003">2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b138">
	<analytic>
		<title level="a" type="main">Texture classification based on support vector machine and wavelet transform</title>
		<author>
			<persName><forename type="first">Bing-Yu</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">De-Shuang</forename><surname>Huang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the Fifth World Congress on Intelligent Control and Automation</title>
				<meeting>of the Fifth World Congress on Intelligent Control and Automation<address><addrLine>WCICA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2004-06-15">2004. June 15-19, 2004</date>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="1862" to="1864" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b139">
	<analytic>
		<title level="a" type="main">Unsupervised texture classification and segmentation</title>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">P</forename><surname>Subramanyam Rallabandi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">K</forename><surname>Sett</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Proceedings Of World Academy of Science, Engineering and Technology</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<date type="published" when="2005-04">April 2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b140">
	<analytic>
		<title level="a" type="main">Use of Gabor filters for texture classification of digital images</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>Recio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">A</forename><surname>Ruiz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Fernández-Sarriá</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Física de la Tierra</title>
		<imprint>
			<biblScope unit="volume">17</biblScope>
			<biblScope unit="page" from="47" to="59" />
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b141">
	<analytic>
		<title level="a" type="main">Texture discrimination by Gabor functions</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">R</forename><surname>Turner</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Biol. Cybern</title>
		<imprint>
			<biblScope unit="volume">55</biblScope>
			<biblScope unit="issue">2-3</biblScope>
			<biblScope unit="page" from="71" to="82" />
			<date type="published" when="1986">1986</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b142">
	<analytic>
		<title level="a" type="main">Texture segmentation using Gabor filters</title>
		<author>
			<persName><forename type="first">V</forename><surname>Levesque</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Center for Intelligent Machines Journal</title>
		<imprint>
			<date type="published" when="2000">2000</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b143">
	<monogr>
		<title level="m" type="main">Segmentation and classification of multi-textured images</title>
		<author>
			<persName><forename type="first">P</forename><surname>Guha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Banerjee</surname></persName>
		</author>
		<ptr target="http://www.cse.iitk.ac.in/~amit/courses/768/00/rajrup/" />
		<imprint>
			<date type="published" when="2000-04-20">2000. April 20, 2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b144">
	<analytic>
		<title level="a" type="main">Automated texture analysis with Gabor filters</title>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">S</forename><surname>Vyas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rege</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">GVIP Journal</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="35" to="41" />
			<date type="published" when="2006-07">July 2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b145">
	<analytic>
		<title level="a" type="main">Wavelets and Support Vector Machines for Texture Classification</title>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">M</forename><surname>Rajpoot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">M</forename><surname>Rajpoot</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">proceedings of the 8 th International Multitopic Conference</title>
				<meeting>the 8 th International Multitopic Conference<address><addrLine>INMIC</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2004-12">2004. Dec. 2004</date>
			<biblScope unit="page" from="328" to="333" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b146">
	<monogr>
		<title level="m" type="main">Optimal Gabor filter design for texture segmentation</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">M</forename><surname>Tsai</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2000">2000</date>
			<publisher>Chung-Li</publisher>
			<pubPlace>, Taiwan</pubPlace>
		</imprint>
		<respStmt>
			<orgName>Machine Vision Lab, Dept. of Ind. Eng. and Mgmt., Yuan-Ze University</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Technical Report</note>
</biblStruct>

<biblStruct xml:id="b147">
	<analytic>
		<title level="a" type="main">Multichannel Texture Analysis Using Localized Spatial Filters</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">C</forename><surname>Bovik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Clark</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">S</forename><surname>Geisler</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Pattern Analysis and Machine Intelligence</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="55" to="73" />
			<date type="published" when="1990">1990</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b148">
	<analytic>
		<title level="a" type="main">On-road vehicle detection using Gabor filters and support vector machines</title>
		<author>
			<persName><forename type="first">Zehang</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Bebis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Miller</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">14 th International Conference on Digital Signal Processing</title>
				<meeting><address><addrLine>DSP</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2002">2002. 2002</date>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="1019" to="1022" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b149">
	<monogr>
		<title level="m" type="main">Texture segmentation using Gabor filters</title>
		<author>
			<persName><forename type="first">K</forename><surname>Hammouda</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Jernigan</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2000-12">Dec. 2000</date>
		</imprint>
		<respStmt>
			<orgName>Biotechnology and health engineering centre, University of Waterloo</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">tech. rep.</note>
</biblStruct>

<biblStruct xml:id="b150">
	<analytic>
		<title level="a" type="main">Comparison of texture features based on Gabor filters</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">E</forename><surname>Grigorescu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Petkov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Kruizinga</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Trans. On Image Processing</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<biblScope unit="issue">10</biblScope>
			<biblScope unit="page" from="1160" to="1167" />
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b151">
	<analytic>
		<title level="a" type="main">Gabor filter based fingerprint classification using support vector machines</title>
		<author>
			<persName><forename type="first">D</forename><surname>Batra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Singhal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Chaudhury</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE First India Annual Conference</title>
				<meeting>the IEEE First India Annual Conference<address><addrLine>INDICON</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2004-12">2004. 2004. Dec. 2004</date>
			<biblScope unit="page" from="20" to="22" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b152">
	<analytic>
		<title level="a" type="main">IRIS Recognition based on multi-channel feature extraction using gabor filters</title>
		<author>
			<persName><forename type="first">Q</forename><forename type="middle">A</forename><surname>Salih</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Dhandapani</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2 nd IASTED international conference on Advances in computer science and technology, ACST&apos;06</title>
				<meeting>the 2 nd IASTED international conference on Advances in computer science and technology, ACST&apos;06</meeting>
		<imprint>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="168" to="173" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b153">
	<analytic>
		<title level="a" type="main">Iris recognition based on multichannel Gabor filtering</title>
		<author>
			<persName><forename type="first">L</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Tan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">5 th Asian Conf. Computer Vision</title>
				<imprint>
			<date type="published" when="2002">2002</date>
			<biblScope unit="volume">1</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b154">
	<analytic>
		<title level="a" type="main">Iris recognition: Gabor filtering</title>
		<author>
			<persName><forename type="first">D</forename><surname>Carr</surname></persName>
		</author>
		<ptr target="http://cnx.org/content/m12493/1.4/" />
	</analytic>
	<monogr>
		<title level="j">Connexions</title>
		<imprint>
			<date type="published" when="2004-12-18">Dec. 18, 2004. April 20, 2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b155">
	<monogr>
		<title level="m" type="main">Feature extraction using Gabor filters</title>
		<author>
			<persName><forename type="first">K</forename><surname>Kämäräinen</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2003-11">Nov. 2003</date>
		</imprint>
		<respStmt>
			<orgName>Lappeenranta University of Technology, Finland</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Ph. D. dissertation</note>
</biblStruct>

<biblStruct xml:id="b156">
	<analytic>
		<title level="a" type="main">Gabor filter design for multiple texture segmentation</title>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">P</forename><surname>Weldon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">E</forename><surname>Higgins</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">F</forename><surname>Dunn</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Optical Engineering</title>
		<imprint>
			<biblScope unit="volume">35</biblScope>
			<biblScope unit="page" from="2852" to="2863" />
			<date type="published" when="1996">1996</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b157">
	<analytic>
		<title level="a" type="main">Evaluation of the effects of Gabor filter parameters on texture classification</title>
		<author>
			<persName><forename type="first">F</forename><surname>Bianconi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Fernández</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Pattern Recognition</title>
		<imprint>
			<biblScope unit="volume">40</biblScope>
			<biblScope unit="issue">12</biblScope>
			<biblScope unit="page" from="3325" to="3335" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b158">
	<monogr>
		<title level="m" type="main">Efficient computation of Gabor features</title>
		<author>
			<persName><forename type="first">J</forename><surname>Ilonen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">K</forename><surname>Kämäräinen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">K</forename><surname>Kälviäinen</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2005">2005</date>
		</imprint>
		<respStmt>
			<orgName>Lappeenranta University of Technology, Dept. of Information Technology</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Research Report 100</note>
</biblStruct>

<biblStruct xml:id="b159">
	<analytic>
		<title level="a" type="main">Las máquinas de vectores de soporte para identificación en línea</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>Reséndiz</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="s">Masters dissertation, Departamento de control automático</title>
		<imprint>
			<date type="published" when="2006">2006</date>
			<publisher>Centro de investigación y estudios avanzados</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b160">
	<monogr>
		<title level="m" type="main">A short SVM (support vector machine) tutorial</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">P</forename><surname>Lewis</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2004">2004</date>
		</imprint>
		<respStmt>
			<orgName>CGIT Lab / IMSC, University Southern California</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b161">
	<analytic>
		<title level="a" type="main">Modelos de clasificación basados en máquinas de vectores de soporte</title>
		<author>
			<persName><forename type="first">L</forename><surname>González</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Asoc. científica europea de econ. aplicada. Anales de economía aplicada</title>
		<imprint>
			<date type="published" when="2003">2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b162">
	<analytic>
		<title level="a" type="main">Texture segmentation using 2-D Gabor elementary functions</title>
		<author>
			<persName><forename type="first">D</forename><surname>Dunn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">E</forename><surname>Higgins</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wakeley</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Pattern Analysis and Machine Intelligence</title>
		<imprint>
			<biblScope unit="volume">16</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="130" to="149" />
			<date type="published" when="1994-02">Feb 1994</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b163">
	<monogr>
		<title level="m" type="main">A complete object oriented environment for machine learning in MATLAB</title>
		<author>
			<persName><surname>Spider</surname></persName>
		</author>
		<ptr target="http://www.kyb.mpg.de/bs/people/spider/" />
		<imprint>
			<date type="published" when="2009-05-15">May 15, 2009</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
