<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Concept Lattice Mining for Unsupervised Named Entity Annotation</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Thomas</forename><surname>Girault</surname></persName>
							<email>thomas.girault@orange-ftgroup.com</email>
							<affiliation key="aff0">
								<orgName type="institution">Orange labs</orgName>
								<address>
									<addrLine>2 avenue Pierre Marzin</addrLine>
									<postCode>22307</postCode>
									<settlement>Lannion Cedex</settlement>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="institution">IRISA</orgName>
								<address>
									<addrLine>Campus universitaire de Beaulieu</addrLine>
									<postCode>35042</postCode>
									<settlement>Rennes Cedex</settlement>
									<country key="FR">France</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Concept Lattice Mining for Unsupervised Named Entity Annotation</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">EA5B96080F7AB1B2E50F963127C9AC79</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T10:36+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>We present an unsupervised method for named entity annotation, based on concept lattice mining. We perform a formal concept analysis from relations between named entities and their syntactic dependencies observed in a training corpus. The resulting lattice contains concepts which are considered as labels for named entities and context annotation. Our approach is validated through a cascade evaluation which shows that supervised named entity classification is improved by using the annotation produced by our unsupervised disambiguation system.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Lexical ambiguity is a fundamental problem which is central in many tasks involving natural language processing (e.g. information retrieval, information extraction, . . . ). Our study focuses on a kind of lexical units (LU), named entities (NE), a generic denomination for proper names including persons, locations, organisations. As most LU considered outside a context, NE are ambiguous since their form can potentially refer to different meanings or objects. Our approach to disambiguation is based on formal concept analysis (FCA), a generic method for data analysis and knowledge representation which infers formal concepts from relational data. In this work, FCA is used to build a knowledge-base that is exploited for NE annotation.</p><p>The problem of ambiguity can be considered according to several Word Sense Disambiguation (WSD) approaches <ref type="bibr">[1]</ref>. Knowledge-based approaches attempt to select the meaning of words using lexicons, dictionaries or thesauri (e.g. Word-Net). Corpus-based approaches examine the occurrence of LU and their contexts using machine learning techniques. Supervised learning disambiguates LU according to pre-defined labels whereas unsupervised techniques discriminate the meanings of unlabelled LU thanks to similarity of their contexts.</p><p>Since corpus annotation is a tedious and costly task, this work is focused on unsupervised approaches. Among them, formal concept analysis (FCA) <ref type="bibr">[2]</ref> has been selected : this symbolic unsupervised machine learning technique operates on relational data to infer formal concepts which can be structured into a concept lattice. FCA is applied on relations between NE and their syntactic dependencies extracted from English news wire articles. The sets of NE sharing the same syntactic dependencies constitute the formal concepts which are considered as units of meaning for the annotation of NE. The concept lattice obtained can be seen as a hierarchical knowledge-base modelling meaning overlapping on several levels of granularity. To our knowledge, these properties attached to concept lattices have not yet been exploited in an unsupervised WSD task. In this context, we propose a conceptual annotation method for NE disambiguation.</p><p>In this paper, we address the problem of exploiting a concept lattice for unsupervised NE annotation. We first introduce (Section 2) the problem of NE ambiguity by exposing few examples from our corpus in which relations between NE and their syntactic dependencies are extracted. These relations constitute a formal context from which FCA is performed (section 3). The resulting lattice contains formal concepts which are considered as labels for NE and dependency annotation (Section 4). Our approach is validated through a cascade evaluation (section 5) which shows that supervised NE classification is improved by using the annotation produced by our unsupervised disambiguation system.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Corpus-Based Methods for Word Sense Disambiguation</head><p>This section introduces corpus-based word sense disambiguation (WSD) with a small sample of a corpus where NE occurrences are semantically labelled. Supervised learning disambiguates LU according to labelled pre-defined meanings whereas unsupervised techniques discriminate the meanings of unlabelled LU thanks to similarity of their lexical contexts. Our unsupervised approach is built upon the study of syntactic relations between NE and other LU occurring in an utterance.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Tagset Granularity for Supervised NE Classification</head><p>Named Entity Recognition (NER) is a subtask of Information Extraction. Different NER systems were evaluated, among others, as a part of the Message Understanding Conferences <ref type="bibr">[3]</ref> in 1995 and in the CoNLL 2003 shared task <ref type="bibr">[4]</ref>. The most efficient NER systems are built upon supervised corpus-based learning for the detection and classification of NE. They rely on semantically annotated corpora which we can illustrate with the following examples (figure 2.1) :</p><p>1. India loc has acquired 120,000 tonnes of diesel in three cargoes, . . . 2. Cricket -: India loc wins the toss and bat against Sri Lanka loc .  The English CoNLL 2003 data is a collection of news wire articles from the Reuters Corpus in which the NE are manually labelled with respect to the coarse-grained semantic tagset {person, location, organisation, miscellaneous}.</p><p>The examples (1) and (2) illustrate a case of ambiguity : the NE "India" is labelled as location but a more fined granularity would distinguish the sport nation and the wholesale importer. In addition we could note that LU interacting with NE are ambiguous as well : the LU "wins" occurs with different meanings for the domains of politics, sport or business. Thus, we think that the original tagset should be enriched with a refined semantic description. However, a manual refinement would be a tedious and a costly task. In addition, we cannot define a general semantic tagset since it is domain dependent : for instance a biomedical semantic tagset should discriminate viruses and proteins and it would not be suitable to describe geographic entities such as rivers or mountains.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Unsupervised Corpus-Based Disambiguation</head><p>Instead of assigning predefined labels to LU, an alternative strategy is to discriminate their meanings by analysing their co-occurrences in the utterances of a corpus. This unsupervised approach is founded from the assumption that LU (NE in our case) which occur in similar contexts tends to have close meanings. Distributional methods <ref type="bibr">[5]</ref> relying on Harris' hypothesis consider that the share of contexts having common syntactic patterns (e.g. subject-verb, modifier-noun) constitutes an indicator of semantic relatedness.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3">Named Entity Dependency Extraction</head><p>Before applying distributional hypothesis for NE disambiguation, the LU attached syntactically to NE need to be identified. We suppose that the NE frontiers have been already detected. Our method deals with two kinds of dependencies. External dependencies are mainly nouns, verbs and prepositions occurring before or after a NE. They are extracted with patterns defined manually relying on morphosyntactic tagging and phrase chunking available with the CoNLL-2003 corpus <ref type="foot" target="#foot_0">1</ref> . The patterns extracts expressions such as :</p><p>noun + preposition + NE (e.g. [election of, Clinton], [results of, European Super League]); noun + NE (e.g. [champion, Pete Sampras]); -NE + noun (e.g. [Russian, government]); -NE + verb (e.g. [Clinton, signed], [India, wins]).</p><p>Internal dependencies correspond to non prepositional tokens occurring in the NE, such as first names or surnames. For example, the list of internal dependencies of International Boxing Federation, is {international, boxing, federation}.</p><p>This work on extraction provides a set of pairs (NE, syntactic dependency) where each element is potentially ambiguous.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Formal Concept Analysis for Knowledge Base Acquisition</head><p>In this section, the approach for knowledge-base acquisition using FCA is exposed. We illustrate FCA with examples taken from our linguistic data. We then discuss the advantages of FCA for dealing with meaning overlapping and granularity of meanings.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Formal Context of Syntactic Relations</head><p>Classical distributional methods could deal with ambiguity of the whole set of LU. However, these methods consider them from a unique point of view whereas for our problem, the data seems more naturally represented according to two interconnected views as the figure <ref type="bibr">(2)</ref> shows :</p><p>a view on named entities which is associated to a set of objects</p><formula xml:id="formula_0">O = {o 1 , o 2 , • • • , o m }.</formula><p>a view on their dependencies (syntactic co-texts + internal components)</p><p>represented by a set of attributes In the FCA terminology, the triple K = (O, A, R) is called a formal context. It corresponds to a bigraph (from the figure (2)) of objects (NE) in relation with attributes (syntactic co-texts + internal components).</p><formula xml:id="formula_1">A = {a 1 , a 2 , • • • , a n }</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Formal Concept Analysis</head><p>For the understanding of the paper, we introduce standard definitions and notations of FCA <ref type="bibr">[2]</ref>. We can define a formal concept of the formal context K to be a pair (E, I) satisfying E ⊆ O, I ⊆ A, E = I and I = E. E is called the extent and I is called the intent of concept. For instance, the pair ({Bill Clinton, Arnold Schwarzenegger},{wins, election of, speech of}) is a formal concept. The concepts are partially ordered according to the relation ≤ :</p><formula xml:id="formula_2">(E 1 , I 1 ) ≤ (E 2 , I 2 ) ⇔ E 1 ⊆ E 2 ⇔ I 2 ⊆ I 1</formula><p>For instance, we have C 2 ≤ C 0 for the concepts C 2 = ({Arnold Schwarzenegger,Roger Federer}, {wins,interview of, fan of}) and C 0 = ({Michael Jackson, Roger Federer, Arnold Schwarzenegger}, {interview of, fan of}). The relation ≤ form a complete lattice L, called the concept lattice of K.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">The Concept Lattice : a Discriminative Knowledge Base</head><p>The general approach for building the concept lattice from linguistic data is similar to the work of Cimiano et al. <ref type="bibr">[6]</ref>. The algorithm AddIntent <ref type="bibr">[7]</ref> has been used for the construction of the lattice. It adopts an incremental procedure allowing dynamic lattice structuring according to new objects or attributes discovered from new utterances. Thus, a lattice could be seen as a knowledge-base already structured which could be adapted to a new corpus. This is an interesting property considering the weak evolutivity of classical lexical resources such as thesauri. According to this perspective, Priss <ref type="bibr">[8]</ref> has been able to encapsulate the FrameNet thesauri within relational concept analysis framework.</p><p>As the figure (3) depicts, the concept lattice structure is organised according to several granularity layers. The upper part of the lattice is represented by general concepts grouping objects which share ambiguous attributes. The opposite part of the lattice has very specific concepts having ambiguous objects. The intermediate zone of the lattice provides concepts which seem more appropriate for LU disambiguation. Although the lattice model is generally considered as symbolic and discrete representation, the intent/extent overlapping reveals potential continuity of meanings. To our knowledge, these properties attached to concept lattices have not been exploited yet for an unsupervised WSD task.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Unsupervised Named Entity Annotation</head><p>In this section, we describe our FCA based methodology for annotation of relations between a NE and its context in an utterance. FCA is not only used to aggregate data, but also to perform a classification of NE. The unsupervised annotation is based on a selection of formal concepts according to n NE and its dependencies. We illustrate the method with an example and we finally introduce a dimensionality reduction method for the visualisation of formal concepts.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Concept Lattice Mining for Conceptual Annotation</head><p>Formal concepts are now considered as units of meaning potentially useful for LU annotation. As we noticed previously, the overlapping of intents and extents between formal concepts is linked to the intuition that some concepts are more similar than others since they share more objects or more attributes. Thus, the formal concepts could be associated to a metric space where the distance between two concepts measures a degree of semantic similarity.</p><p>In a new utterance, we suppose that a new NE o ∈ O and its dependencies Atts ⊆ A have been detected thanks to the extraction patterns (section 2.3). For a disambiguation task, we consider that the meaning of o relies on the meaning of its dependencies in Atts occurring in the context : in other words, o can be annotated with a formal concept x ∈ L according to the concepts for the dependencies in Atts.</p><p>Our model for conceptual annotation of named entities is based on querying the concept lattice. In the lattice L the object o is associated to Co = ({o} , {o} ) and similarly the concepts for the attributes of Atts are the elements Ca i from C Atts = {({a i } , {a i } )|a i ∈ Atts}. We are looking for a representative concept in the lattice which interpolates the concepts Co and Ca i . We will call this concept x the prototype and we search it among the concepts containing o in their extent or at least one dependency a i in their intent. More formally, x ∈ L(o, Atts) where L(o, Atts) = {(E, I) ∈ L|o ∈ E ∨ Atts ∩ I = ∅}. The prototype x is defined as the concept whose average dissimilarity to the concepts Co and Ca i is minimal.</p><formula xml:id="formula_3">X = argmin x∈L(o,Atts) c∈C Atts ∪{Co} similarity(c, x)<label>(1)</label></formula><p>In order to deal with similarities, we define two matrices A(o, Atts) and O(o, Atts) in which each row corresponds to a formal concept from L(o, Atts). The columns of A(o, Atts) are assigned to the intent of the concepts and similarly, the columns of O(o, Atts) are assigned to the extent of the concepts. Thus, the formal concepts are represented by a vector for extents and a vector for intents. Note that we can also consider M(o, Atts) which is the concatenation of the matrices A(o, Atts) and O(o, Atts).</p><p>Similarity measures can then be applied between the concept vectors of A(o, Atts), O(o, Atts) or M(o, Atts) : measures such as Euclidean, cosine, correlation, Hamming or Jaccard can be chosen, depending of if we consider the vectors (and the formal context) as boolean or as weighted by the frequency counts of relations (cooccurrences) observed in the corpus. In the last case, the weights assigned to objects and attributes would be respectively </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Example from CoNLL Data</head><p>To illustrate the method, we propose to annotate the expression "English division" from which the pair (o, Atts) = (English,{division}) is extracted. In a classical dictionary, the LU division is typically ambiguous because it can denotes, for instance, a group of military troops or a group of teams in an organised sport. The following list enumerates the concepts associated to (o, Atts)</p><p>In the lattice L(English,{division}), the object "English" is represented in the lattice by the concept C 9 and the attribute "division" is represented by the concept C 5 . Most concepts appears to denote the sport division meaning and it remains to select an appropriate concept for the annotation of the query. The prototype calculation has been done on this example according to several similarity metrics. The Euclidean and hamming distances chosen among others for the similarity measures, have both selected the concept C 4 which seems a acceptable for the annotation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">Dimensionality Reduction for Visualisation of Formal Concepts</head><p>The technique presented here has not yet been linked to the disambiguation process. It illustrates our intuition that continuous semantic provided with distance fits with a high structured representation such as concept lattices. For a better understanding of this intuition, we propose to visualise formal concepts through a cartographic representation where distance between formal concepts translates the notion of semantic proximity.</p><p>We have describe previously a simple way to associate a set of formal concepts to matrices. Since the vectors associated to concepts potentially have a huge dimension, we propose to use dimensionality reduction methods on the matrix M(o, Atts). These methods are able to compress M(o, Atts) such as each vector/concept representation is reduced to two dimensions. Among these methods we have chosen curvilinear component analysis (CCA) <ref type="bibr">[9]</ref> which can be seen as a non linear extension to principal component analysis. The first results of this method are depicted by the figure (4). Reduced labelling has been used to improve the readability of the figure. In this scheme, the label for an object o is drawn above the object concept γ(o) = ({o} , {o} ) while the label for an attribute a is drawn below the attribute concept µ(a) = ({a} , {a} ).</p><p>Our approach does not take advantage of the partial ordering between concepts that has been already computed. However, according to these figure, the general to specific ordering seems globally respected whereas it has not been taken into account for the rendering of the map : the most general and the most specific concepts occur to opposite sides of the map. The figure (4) also helps to understand where is the prototype C 4 among the other concepts resulting to the query.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Experiments and Evaluation</head><p>Previously, we have described an unsupervised method for conceptual annotation of NE. The evaluation of such unsupervised methods is subjective by nature since several concepts would be relevant to disambiguate a NE. In this section, we present a validation of our approach according to an existing task (supervised NE classification) that we are able to evaluate the performance. We then describe the cascade evaluation protocol <ref type="bibr">[10]</ref> which considers the unsupervised conceptual annotation as a pre-processing step for a supervised NE classification task. We conclude the section with a study of the results obtained through this experiment.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1">CoNLL 2003 Data</head><p>The CoNLL-2003 named entity English data consists of three files : one training file (train), one development file (testa) and one test file (testb). Figure <ref type="bibr">(5)</ref> gives an overview of the characteristics of the corpus.  The Euclidean measure has been used for the prototype determination of the intent matrix A(Lindsay Davenport, {champion}) and for the extent matrix O(Lindsay Davenport, {champion}). It selects two concepts C 52 and C 46 : the intent of C 52 provides a disambiguation of "champion" and the extent of C 46 gives an annotation for "Lindsay Davenport".</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2">Cascade Evaluation</head><p>In the framework of cascade evaluation <ref type="bibr">[10]</ref>, unsupervised learning is considered as a pre-processing step for a supervised NE classification task that we are able to evaluate. This cascade process reveals whether the conceptual annotation provides interesting enrichments to improve the supervised task on the CoNLL 2003 corpus. The protocol consists in comparing errors produced by two classifiers A and B, when they perform on the test corpus (testb), after a training step on the same training data (train + testa).</p><p>The system A is a supervised classifier trained normally on the labelled training corpus. As Ehrmann and Jacquet proposed <ref type="bibr">[11]</ref>, the system B provides two annotations for NE. The first is given by our unsupervised annotation system exploiting the concept lattice learned on the unlabelled training corpus. This pre-processing step provides enrichments to the initial corpus description. The system B can then benefit from these additional enrichments during the supervised learning step in order to produce the second annotation layer.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3">Experimental Results with Transformation-Based Learning</head><p>We have adapted the transformation-based learning (TBL) algorithm <ref type="bibr">[12]</ref> to design a supervised NER system. The algorithm initializes the NE labels with a language model classifier (unigram), trained on the training corpus. The goal is to correct this initial classification according to the original NE labels specified in the training corpus. The next steps follow an iterative process : it corrects the initial incorrect classification by inferring a sequence of transformation rules. They are successively applied over the corpus in order to improve progressively the NE classification.</p><p>The resulting rules are instantiated from a list of extraction patterns defined manually. These patterns are able to explore co-texts features in a window of +/-3 words : among the available features, we have considered the word, its morphosyntactic tag and the concept identifiers given by our unsupervised conceptual annotation method.</p><p>The figure <ref type="bibr">(7)</ref> shows the results of the cascade evaluation. The left column indicates the performances reached by classifier A applied on the test corpus provided with morphosyntactic tagging. The right column corresponds to results obtained with the classifier B which has be used on the test corpus enriched with the conceptual annotation.  According to these results, the unsupervised annotation system increases the precision score to 11.14% and the F β=1 (where F β = (1+β 2 )•(precision•recall)</p><formula xml:id="formula_4">β 2 •precision+recall</formula><p>) measure to 2.61. However, a regression of 4.4% has been observed for recall.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Conclusion, Discussion and Future Work</head><p>We have presented an unsupervised method for named entity annotation, which is based on formal concept analysis. This method exploits a concept lattice structuring relations between named entities and their related lexical units, observed in text corpora. We have assumed that formal concepts are relevant units for the disambiguation of named entities. The selection of a concept for an annotation results of a query to the lattice. In addition, we have proposed a method based on dimensionality reduction for the visualisation of formal concepts. We have adapted the cascade evaluation protocol to validate the choice of concepts for annotation. It shows that a supervised named entity classifier improves its precision when it relies on the conceptual annotation produced by our unsupervised FCA-based system. Even if, our system does not reach the performances obtained by the best named entity recognizers, the first results are encouraging since some improvements are possible.</p><p>The syntactic extraction process could be improved by using a dependency parser : this could help to cover more syntactic patterns. It could also provide some additional information such as normalised forms (e.g. {is, was, were} → to be) or typed syntactic relations (e.g. subject-object, head-modifier).</p><p>The cascade evaluation framework, could compare our approach to other supervised and unsupervised classifiers : we would be particularly interested in the comparison with other FCA based classifiers <ref type="bibr">[13]</ref>. At the present time, we are working on a semi-supervised lattice based classifier in which formal concepts are tagged with the NE labels (persons, locations, organisations, miscellaneous) available in the training corpus. Thus, the lattice would then be usable directly as a supervised NE classifier which would be able to produce unsupervised conceptual annotation with additional supervised labelling.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Samples extracted from the English CoNLL-2003 annotated corpus.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. Relations between NE and their dependencies.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head></head><label></label><figDesc>For E ⊆ O and I ⊆ A, we define two sets E ⊆ A and I ⊆ O extending them : E = {a ∈ A|∀o ∈ E : (o, a) ∈ R} as the set of all attributes from I that are in relation with all objects from E and I = {o ∈ O|∀a ∈ I : (o, a) ∈ R}, the set of all objects from O that are in relation with all attributes from I. For instance, if I = {speech of, election of} then I = {Bill Clinton, Arnold Schwarzenegger}. For E = {Michael Jackson}, we have, E = {album, live performance of, interview of, fan of}.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>c 4 6</head><label>46</label><figDesc>Michael album of,interview of, live performance of,fan of c Arnold actor, fan of, film with, wins, election of, speech of ⊥ ∅ wins, election of, speech of, interview of, fan of, live performance, album of, actor, film with, match against</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Fig. 3 .</head><label>3</label><figDesc>Fig.3. Concept lattice for the formal context of figure(2). A concept box is contains a name, an extent and an intent.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head></head><label></label><figDesc>ai∈intent(C) card(R(o, a i )) and oi∈extent(C) card(R(o i , a)).</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head>Fig. 4 .</head><label>4</label><figDesc>Fig. 4. Visualisation of formal concepts associated to the query (English, {division}) using CCA.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_7"><head>Fig. 6 .</head><label>6</label><figDesc>Fig. 6. Example of conceptual annotation in the CoNLL 2003 corpus.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_8"><head>Fig. 7 .</head><label>7</label><figDesc>Fig. 7. Cascade evaluation results.</figDesc></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">Morphosyntactic tagging and chunking have been generated automatically and are therefore noisy.Concept Lattice Mining for Unsupervised Named Entity Annotation</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" xml:id="foot_1">Thomas Girault   </note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" xml:id="foot_2">Concept Lattice Mining for Unsupervised Named Entity Annotation</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">ENGLISH&apos;</title>
	</analytic>
	<monogr>
		<title level="m">SCOTTISH PREMIER DIVISION</title>
				<imprint/>
	</monogr>
	<note>division&apos;. &apos;premier</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">ENGLISH&apos;</title>
	</analytic>
	<monogr>
		<title level="m">division&apos;, &apos;results</title>
				<imprint/>
	</monogr>
	<note>DUTCH&apos;. SCOTTISH&apos;. &apos;league</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">ENGLISH</title>
	</analytic>
	<monogr>
		<title level="m">division&apos;, &apos;draw&apos;, &apos;division leaders&apos;, &apos;league&apos;, &apos;results&apos;, &apos;result</title>
				<imprint/>
	</monogr>
	<note>DUTCH&apos;. news agency</note>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">ENGLISH</title>
	</analytic>
	<monogr>
		<title level="m">league soccer&apos;, &apos;league&apos;, &apos;division&apos;, &apos;premier&apos;, &apos;results&apos;, &apos;league standings&apos;, &apos;league cup&apos;</title>
				<imprint/>
	</monogr>
	<note>SCOTTISH&apos;. &apos;summaries</note>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m">SCOTTISH PREMIER DIVISION</title>
				<imprint/>
	</monogr>
	<note>HUNGARIAN&apos;. division&apos;</note>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">ENGLISH&apos;</title>
	</analytic>
	<monogr>
		<title level="m">division&apos;, &apos;referee&apos;</title>
				<imprint/>
	</monogr>
	<note>WELSH. &apos;results</note>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">GERMAN&apos;</title>
	</analytic>
	<monogr>
		<title level="m">division&apos;, &apos;results</title>
				<imprint/>
	</monogr>
	<note>DUTCH&apos;. ENGLISH&apos;. &apos;result</note>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">ENGLISH&apos;</title>
	</analytic>
	<monogr>
		<title level="m">result&apos;, &apos;division&apos;, &apos;law&apos;, &apos;summaries</title>
				<imprint/>
	</monogr>
	<note>GERMAN&apos;. &apos;results</note>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<title level="m">standings&apos;, &apos;premier&apos;, &apos;results&apos;, &apos;county&apos;, &apos;result&apos;, &apos;league cup&apos;, &apos;news agency&apos;, &apos;city&apos;, &apos;langage&apos;, &apos;soccer matches&apos;</title>
				<imprint/>
	</monogr>
	<note>play scores</note>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">SCOTTISH&apos;</title>
	</analytic>
	<monogr>
		<title level="m">AUSTRIA&apos;, &apos;DUTCH&apos;, &apos;MOROCCAN&apos;, &apos;ENGLISH&apos;, &apos;SWISS&apos;</title>
				<imprint/>
	</monogr>
	<note>division&apos;. &apos;results</note>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">ENGLISH</title>
	</analytic>
	<monogr>
		<title level="m">division&apos;, &apos;premier&apos;, &apos;results</title>
				<imprint/>
	</monogr>
	<note>FRENCH&apos;. SCOTTISH&apos;. &apos;summaries</note>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m">division&apos;, &apos;draw&apos;</title>
				<imprint/>
	</monogr>
	<note>AUSTRIA&apos;. &apos;results</note>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<title level="m" type="main">division&apos;, &apos;results</title>
		<imprint/>
	</monogr>
	<note>ENGLISH&apos;. &apos;league leaders</note>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<title level="m">DUTCH&apos;, &apos;ENGLISH&apos;, &apos;AMERICAN LEAGUE EAST DIVISION&apos;, &apos;NATIONAL LEAGUE CENTRAL DIVISION&apos;</title>
				<imprint/>
	</monogr>
	<note>NATIONAL LEAGUE EASTERN DIVISION. SCOTTISH&apos;. &apos;league&apos;, &apos;division</note>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">ENGLISH&apos;</title>
	</analytic>
	<monogr>
		<title level="m">SCOTTISH PREMIER DIVISION</title>
				<imprint/>
	</monogr>
	<note>SCOTTISH PREMIER</note>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<title level="m">&apos;division leaders</title>
				<imprint/>
	</monogr>
	<note>news agency. SCOTTISH</note>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<title level="m">&apos;league soccer&apos;, &apos;league standings&apos;</title>
				<imprint/>
	</monogr>
	<note>&apos;league cup</note>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<title/>
		<author>
			<persName><surname>Welsh</surname></persName>
		</author>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<title level="m">&apos;standings&apos;, &apos;county championship matches&apos;, &apos;county&apos;, &apos;city&apos;, &apos;langage&apos;, &apos;soccer matches</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<title level="m" type="main">MOROCCAN</title>
		<imprint/>
	</monogr>
	<note>POLISH&apos;</note>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">&apos;results</title>
	</analytic>
	<monogr>
		<title level="m">NATIONAL LEAGUE EASTERN DIVISION</title>
				<imprint>
			<biblScope unit="volume">14</biblScope>
		</imprint>
	</monogr>
	<note>NATIONAL LEAGUE CENTRAL DIVISION</note>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title/>
	</analytic>
	<monogr>
		<title level="j">AUSTRIA</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<monogr>
		<title level="m">DIVISION&apos;, &apos;ROMANIA&apos;, &apos;FIRST UNION CORP&apos;, &apos;HUNGARIAN&apos;, &apos;TAICHUNG DIVISION</title>
				<imprint/>
	</monogr>
	<note>FRASER&apos;</note>
</biblStruct>

<biblStruct xml:id="b23">
	<monogr>
		<title level="m" type="main">In: Unsupervised corpus-based methods for WSD</title>
		<author>
			<persName><forename type="first">T</forename><surname>Pedersen</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2006">2006</date>
			<publisher>Springer</publisher>
			<biblScope unit="page" from="133" to="166" />
		</imprint>
	</monogr>
	<note>6</note>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Formal Concept Analysis</title>
		<author>
			<persName><forename type="first">B</forename><surname>Ganter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Wille</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Mathematical Foundations</title>
				<imprint>
			<publisher>Springer-Verlag</publisher>
			<date type="published" when="1999">1999</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Message understanding conference 6: A brief history</title>
		<author>
			<persName><forename type="first">R</forename><surname>Grishman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Sundheim</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">COLING</title>
				<imprint>
			<date type="published" when="1996">1996</date>
			<biblScope unit="page" from="466" to="471" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<monogr>
		<title level="m" type="main">Introduction to the conll-2003 shared task: Language-independent named entity recognition</title>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">F T K</forename><surname>Sang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">D</forename><surname>Meulder</surname></persName>
		</author>
		<idno>CoRR cs.CL/0306050</idno>
		<imprint>
			<date type="published" when="2003">2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<monogr>
		<title level="m" type="main">Corpus-based identification and refinement of semantic classes</title>
		<author>
			<persName><forename type="first">A</forename><surname>Nazarenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Zweigenbaum</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Bouaud</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Habert</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1997">1997</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">Learning concept hierarchies from text corpora using formal concept analysis</title>
		<author>
			<persName><forename type="first">P</forename><surname>Cimiano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hotho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Staab</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Artificial Intelligence Research</title>
		<imprint>
			<biblScope unit="volume">24</biblScope>
			<biblScope unit="page" from="305" to="339" />
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<monogr>
		<title level="m" type="main">Addintent: A new incremental algorithm for constructing concept lattices</title>
		<author>
			<persName><forename type="first">D</forename><surname>Van Der Merwe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">A</forename><surname>Obiedkov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">G</forename><surname>Kourie</surname></persName>
		</author>
		<imprint>
			<biblScope unit="volume">15</biblScope>
			<biblScope unit="page" from="372" to="385" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<analytic>
		<title level="a" type="main">Modelling lexical databases with formal concept analysis</title>
		<author>
			<persName><forename type="first">U</forename><surname>Priss</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">J</forename><surname>Old</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">J. UCS</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="issue">8</biblScope>
			<biblScope unit="page" from="967" to="984" />
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b31">
	<analytic>
		<title level="a" type="main">Curvilinear component analysis for high-dimensional data representation: I. theoretical aspects and practical use in the presence of noise</title>
		<author>
			<persName><forename type="first">J</forename><surname>Hérault</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Jausions-Picaud</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Guérin-Dugué</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IWANN</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="625" to="634" />
			<date type="published" when="1999">1999</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b32">
	<analytic>
		<title level="a" type="main">Cascade evaluation of clustering algorithms</title>
		<author>
			<persName><forename type="first">L</forename><surname>Candillier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Tellier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Torre</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Bousquet</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">17th European Conference on Machine Learning (ECML&apos;2006)</title>
		<title level="s">LNAI</title>
		<editor>
			<persName><forename type="first">J</forename><surname>Fürnkranz</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><surname>Scheffer</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Spiliopoulou</surname></persName>
		</editor>
		<meeting><address><addrLine>Berlin, Germany</addrLine></address></meeting>
		<imprint>
			<publisher>Springer Verlag</publisher>
			<date type="published" when="2006-09">september 2006</date>
			<biblScope unit="volume">4212</biblScope>
			<biblScope unit="page" from="574" to="581" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b33">
	<analytic>
		<title level="a" type="main">Vers une double annotation des entités nommées</title>
		<author>
			<persName><forename type="first">M</forename><surname>Ehrmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Jacquet</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Traitement automatique des langues</title>
		<imprint>
			<biblScope unit="volume">47</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="63" to="88" />
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b34">
	<analytic>
		<title level="a" type="main">Transformation-based error-driven learning and natural language processing: A case study in part-of-speech tagging</title>
		<author>
			<persName><forename type="first">E</forename><surname>Brill</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computational Linguistics</title>
		<imprint>
			<biblScope unit="volume">21</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="543" to="565" />
			<date type="published" when="1995">1995</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b35">
	<monogr>
		<title level="m" type="main">Machine learning and formal concept analysis</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">O</forename><surname>Kuznetsov</surname></persName>
		</author>
		<imprint>
			<biblScope unit="volume">15</biblScope>
			<biblScope unit="page" from="287" to="312" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b36">
	<analytic>
		<title level="a" type="main">Exploitation de treillis de Galois en désambiguïsation non supervisée d&apos;entités nommées</title>
		<author>
			<persName><forename type="first">T</forename><surname>Girault</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">conférence sur le Traitement Automatique des Langues Naturelles (TALN&apos;</title>
		<imprint>
			<biblScope unit="volume">08</biblScope>
			<biblScope unit="page" from="260" to="269" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b37">
	<analytic>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">W</forename><surname>Eklund</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Concept Lattices, Second International Conference on Formal Concept Analysis, ICFCA 2004</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">P</forename><forename type="middle">W</forename><surname>Eklund</surname></persName>
		</editor>
		<meeting><address><addrLine>Sydney, Australia</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2004">February 23-26, 2004. 2961. 2004</date>
		</imprint>
	</monogr>
	<note>ICFCA.</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
