<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">ANEA: Automated (Named) Entity Annotation for German Domain-Specific Texts</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<affiliation key="aff0">
								<orgName type="institution">University of Wuppertal</orgName>
								<address>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff1">
								<orgName type="institution">Felix Hamborg</orgName>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff2">
								<orgName type="institution">University of Konstanz Gipp</orgName>
								<address>
									<settlement>Bela</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff3">
								<orgName type="institution">University of Wuppertal</orgName>
								<address>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">ANEA: Automated (Named) Entity Annotation for German Domain-Specific Texts</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">95C5DE98C83D548BD6D351741B081FF9</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T15:20+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>CCS CONCEPTS</term>
					<term>Information systems → Information extraction</term>
					<term>• Computing methodologies → Information extraction</term>
					<term>Language resources</term>
					<term>Cluster analysis information extraction, low-resource languages, named entity recognition, domain-specific texts</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Named entity recognition (NER) is an important task that aims to resolve universal categories of named entities, e.g., persons, locations, organizations, and times. Despite its common and viable use in many use cases, NER is barely applicable in domains where general categories are suboptimal, such as engineering or medicine. To facilitate NER of domain-specific types, we propose ANEA, an automated (named) entity annotator to assist human annotators in creating domain-specific NER corpora for German text collections when given a set of domain-specific texts. In our evaluation, we find that ANEA automatically identifies terms that best represent the texts' content, identifies groups of coherent terms, and extracts and assigns descriptive labels to these groups, i.e., annotates text datasets into the domain (named) entities.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">INTRODUCTION</head><p>Named entity recognition (NER), a common preprocessing step in natural language processing (NLP) for various tasks, such as information extraction, summarization, question and answering, and text understanding, is often criticized for capably representing only datasets with few general categories, e.g., person, location, organization, and time (including their subcategories) <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b29">30]</ref>. While the original NER task contains only few categories, a rapidly increasing number of NER applications show a high demand for the datasets with domain-specific named entities <ref type="bibr" target="#b17">[18]</ref>.</p><p>To (semi-)automatically create large general-purpose NER corpora, recent research projects extensively use structured domain sources, such as dictionaries, knowledge graphs, and Wikipedia or other knowledge bases <ref type="bibr" target="#b16">[17,</ref><ref type="bibr" target="#b23">24,</ref><ref type="bibr" target="#b30">31]</ref>.</p><p>In this paper, we propose ANEA, an unsupervised Wiktionarybased approach that automatically derives domain entities from German texts, i.e., low-resource language, by (1) extracting terms from topically-related domain texts, (2) identifying the most domainrepresentative, i.e., semantically distinct, terms of the analyzed texts, and (3) automatically annotating the terms, i.e., ANEA extracts labels from Wiktionary and assigns them to the identified groups of terms. Not all of the domain categories may be named, e.g., machinery or process.</p><p>By automating the most labor-intense parts, the proposed unsupervised approach minimizes the cost of expensive and laborious annotations required for the creation of domain-specific NER datasets. Typical manual tasks in annotations include <ref type="bibr" target="#b0">(1)</ref> reading the domain text multiple times, <ref type="bibr" target="#b1">(2)</ref> deriving entities based on the text content, and (3) manually selecting terms that match the derived categories. ANEA substitutes the most time-consuming task of deriving a coding book and automatically defines categories and annotates the most representative terms (nouns) into these categories. We evaluate the approach with user studies on multiple domain datasets against multiple silver datasets and discuss a default input configuration for ANEA to annotate other domain NER datasets 1 .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">RELATED WORK</head><p>NER datasets usually contain standard types, e.g., person, location, organization, and are manually annotated <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b20">21]</ref> or automatically extracted <ref type="bibr" target="#b16">[17,</ref><ref type="bibr" target="#b23">24,</ref><ref type="bibr" target="#b30">31]</ref>. Domain-specific NER typically needs to introduce domain-specific (sub-)categories of the established named entity (NE) categories or entirely new categories. This is because domain-specific texts contain NE categories that are (1) detailed variants of the standard NE categories, e.g., "Person" is replaced with the domain-specific sub-categories "Players" and "Coaches" <ref type="bibr" target="#b26">[27]</ref>, <ref type="bibr" target="#b1">(2)</ref> standard NE categories extended with a small number of new categories, e.g., "Trigger of a traffic jam" <ref type="bibr" target="#b10">[11,</ref><ref type="bibr" target="#b18">19,</ref><ref type="bibr" target="#b21">22]</ref>, and (3) domain-derived NE categories, e.g., "Proteins" in biology or "Reactions" in chemistry domains <ref type="bibr" target="#b8">[9,</ref><ref type="bibr" target="#b17">18,</ref><ref type="bibr" target="#b24">25,</ref><ref type="bibr" target="#b29">30]</ref>. Most domain-derived NE categories originate from structured classifications or dictionaries <ref type="bibr" target="#b8">[9,</ref><ref type="bibr" target="#b11">12,</ref><ref type="bibr" target="#b24">25]</ref> or are derived by manually unifying multiple of them <ref type="bibr" target="#b4">[5]</ref>. In sum, creating domain-specific datasets for NER requires expert knowledge and is time-consuming.</p><p>To minimize such efforts, some NER approaches use seed-NEs, i.e., a small number of manually provided terms and their NEcategories <ref type="bibr" target="#b6">[7,</ref><ref type="bibr" target="#b15">16,</ref><ref type="bibr" target="#b29">30]</ref>. Such approaches use the seed-NEs as examples to extract patterns of NE definitions and apply them to the full text suggested for annotation. These NER approaches suffer from the slow updates of the underlying domain knowledge bases (KB) <ref type="bibr" target="#b27">[28]</ref> and perform worse on lower-resource languages than on English <ref type="bibr" target="#b8">[9]</ref>. An alternative to domain-KBs is community KBs, such as Wikipedia and Wiktionary, which are constantly updated by their communities. They prove to contain a sufficient amount of domain information <ref type="bibr" target="#b13">[14,</ref><ref type="bibr" target="#b28">29]</ref>.</p><p>Unlike the existing supervised approaches for annotating domainspecific named entities <ref type="bibr" target="#b5">[6,</ref><ref type="bibr" target="#b12">13]</ref>, in this paper, we explore ANEA, an unsupervised method to support researchers and users during the creation of a coding book. Given a set of domain-or use case-specific documents, ANEA automatically derives domain-specific categories and exemplary terms within. This way, ANEA automates the most time-intensive, previously manual tasks. As a consequence, users only need to revise these terms, e.g., by renaming the categories or re-annotating the not matching to the categories terms.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">METHODOLOGY</head><p>We propose an unsupervised approach for annotation of domainspecific (named) entities (ANEA) on a lower-resource language. The goal of ANEA is to fully automatically derive entity categories (later in the text: categories) by selecting groups of related terms and extract and assign a meaningful label to these terms. To do so, ANEA, first, links terms extracted from domain-specific texts to pages in Wiktionary <ref type="bibr" target="#b13">[14,</ref><ref type="bibr" target="#b23">24]</ref>. Second, ANEA automatically identifies groups of related terms and automatically labels them by performing a double optimization task of both maximizations of cross-similarity of terms in a group and the average similarity of these terms to a candidate label. That is, the approach consists of two main steps: (1) text preprocessing, i.e., term mapping to Wiktionary pages (WPs) and construction of a domain graph, and (2) identification of related terms and label assignment.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Preprocessing and domain graph</head><p>3.1.1 Preprocessing. The goal of preprocessing is to extract terms from the set of texts and maximize the number of terms aligned to the Wiktionary structure, i.e., map the domain-specific terminology to the structured knowledge base. The mapping of the extracted terms to the knowledge graph enables using their semantic information, such as term definitions, areas, hypernyms, and hyponyms (see Figure <ref type="figure" target="#fig_0">1</ref>). The preprocessing steps include parsing and part-of-speech (POS) tagging using spaCy <ref type="bibr" target="#b9">[10]</ref>. We define a term as any unique noun phrase that does not contain any digits <ref type="bibr" target="#b6">[7]</ref>. After extraction, terms are mapped to their respective German WP, if any, i.e., each term gets assigned to a link of a WP.</p><p>In German texts, we find that many domain-specific terms are compound words, i.e., words that consist of more than one noun component, for example, "Sechszylindermotor" = "sechs" + "Zylinder" + "Motor" (six-cylinder motor). Typically, such complex domain compound words are not described in Wiktionary since they are too rare or specific. On the contrary, we observe that compound words' heads, i.e., the part of a composition term that bears the core meaning of the phrase, e.g., "Motor" in the example above, are highly likely to have a WP in Wiktionary.</p><p>To map rare domain-specific terms to WPs, we extract the heads of the extracted terms with a compound splitter, i.e., a model that splits terms into two parts, compound-and head-parts, <ref type="bibr" target="#b25">[26]</ref> and attempt to map the heads to Wiktionary. If a (multi-token) term has a corresponding WP, we set a full term as a term's head. If the compound splitter outputs a head that is not a part of Wiktionary, we continue recursively search for a head that can be mapped to a WP. If no heads have corresponding WPs, then we do not assign a head to a term.</p><p>The preprocessing could be changed to include the terms with digits, but for now, we focused on the noun phrases as terms. If a term or its head of compound phrase do not have a WP, they are excluded from annotation because the absence of a link to Wiktionary leads to an inability to map terms to potential category labels. Later, such discarded terms can be manually classified by human annotators or automatically with state-of-the-art NER models trained on the automatically created domain datasets.</p><p>We use fastText to vectorize extracted terms and candidate labels <ref type="bibr" target="#b7">[8]</ref>. We chose fastText due to its ability to vectorize out-ofvocabulary words, which often happens with domain-specific terminology.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.2">Domain graph.</head><p>The domain graph is a locally stored knowledge graph where the leaves are the extracted domain terms. Nodes are all terms obtained from the WPs linked to the leaves and to each other with hyponymy-hypernymy relations. Figure <ref type="figure" target="#fig_1">2a</ref> depicts the principle of a domain graph. The construction of the domain graph includes three steps: (1) graph initialization, i.e., extraction of the WP properties, e.g., definitions, with the scraping of the WPs assigned to the domain terms and their heads; (2) determination of pruning criteria of Wiktionary graph to scrape only domain related pages; (3) expansion of the domain graph, i.e., scrapping of the hypernym pages, to create a pool of candidate labels to later annotated the identified groups of terms. Figure <ref type="figure" target="#fig_1">2b</ref> shows the process of domain graph construction.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.3">Initialization.</head><p>To initialize the graph, we use the extracted terms to which we mapped WPs and scrap the mapped WPs to extract WPs' properties, e.g., hypernyms. As a preliminary step of the graph initialization, we group the extracted terms by their head. The head grouping aims at the extraction of the initial hyponymhypernym relation for the domain graph. Then, we sort the list of heads in decreasing order by <ref type="bibr" target="#b0">(1)</ref> the number of unique terms with each head, (2) the frequency of the overall in-text occurrence of words with such head. To maximize the descriptiveness and generalization of the terms that will become annotated into categories, we initialize the domain graph with the terms-to-annotate that belong to the top 𝑀 largest head groups, i.e., containing the most lexically diverse and/or frequent terms. Section 4 determines an optimal value of terms-toannotate and the largest head groups the series of experiments. This filtering procedure reduces the size of the domain graph to minimize the time of the execution and extract the most representative candidate labels, i.e., the most closely located hypernyms.</p><p>Each term without a hyponym is a leaf of the domain graph; a node is a head that aggregates more than one term. We scrape WPs of all leave-and node-terms to extract the text and links from definitions, hypernyms and hyponyms (see Figure <ref type="figure" target="#fig_0">1</ref>).</p><p>We extract hyponym terms from the corresponding WP's section. We extract hypernym terms from two WP's parts: (1) the hypernym section, (2) the definition section by parsing the text of term's definitions and ensuring that the extracted word has its WP <ref type="foot" target="#foot_0">2</ref> . For example, in Figure <ref type="figure" target="#fig_0">1</ref>, the word "Maschine" will be extracted as an additional (in-text) hypernym to those listed in the hypernym section.</p><p>The extracted properties are assigned to each node. The hypernyms' links point at the WPs that may later become nodes of the domain graph. Extraction and assignment of the WP's properties bridge the domain terms and heads to the Wiktionary's knowledge graph.</p><p>3.1.4 A priori pruning. Most of the terms of WPs have more than one sense and some of them may be associated with the different semantic areas, e.g., technology, medicine, sport, law, etc. If a sense belongs to only one area, the title of the area precedes the definition explaining the sense, e.g., "Technik" in Figure <ref type="figure" target="#fig_0">1</ref>. The step a priori pruning determines which senses of yet to add hypernyms need to comply with the senses of the previously added terms. Hypernyms will become properties of a node in the domain graph if and only if their areas belong to a predefined list of areas or if they do not list any domain areas.</p><p>To identify which Wiktionary's areas determine the graph's domain, we select the most frequent and semantically similar areas extracted from the senses' definitions of the previously added leaves and nodes. To find the most semantically similar and frequent areas, we cluster all titles using hierarchical clustering and select the areas from the most representative clusters. As parameters of the hierarchical clustering, we use Euclidean distance, average linkage criterion, and optimize the number of clusters. To represent areas' titles in the vector space, we apply the fastText word embeddings model <ref type="bibr" target="#b7">[8]</ref>.</p><p>We extract three most representative clusters by: (1) selecting a cluster with the average cosine similarity across all words in a cluster being the highest among all clusters (𝐶 𝑠 ); (2) selecting a cluster with the count of all words in a cluster is the highest among all clusters (𝐶 𝑓 ); (3) forming an extra cluster with the 𝐾 most frequent areas 𝐶 𝑘 (𝐾 = 5). To identify the Wiktionary areas 𝐴 forming the domain of the graph, we intersect all three representative clusters: 𝐴 = {∪𝑎 : 𝑎 ∈ 𝐶 𝑠 ∩ 𝐶 𝑓 ∩ 𝐶 𝑘 }. Finally, we select a clustering configuration that outputs the best domain-defining areas as 𝐴 𝑏𝑒𝑠𝑡 = arg max 6≤𝑖 ≤12 (𝑠 (𝐴 𝑖 ) • 𝑓 (𝐴 𝑖 )), where 𝐴 𝑖 are the areas identifies at 𝑖 the number of clusters, 𝑠 (𝐴 𝑖 ) is an average cross similarity of the areas in 𝐴 𝑖 and 𝑓 (𝐴 𝑖 ) is sum of 𝐴 𝑖 ' frequencies.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.5">Graph growing.</head><p>The goal of NEA is to assign the most generalizing yet still representative labels to groups of semantically related terms, e.g., "Person" for "Trump" and "Einstein. " To ensure the generalization property of label candidates, we "grow", i.e., expand the domain graph up, by adding new nodes on the top of the graph from the scrapped hypernym WPs.</p><p>To grow the graph, we iterate over the top nodes and create new nodes for each of the hypernym terms. To obtain node's properties, we scrape WPs of the hypernym terms and extract term definitions, hyponyms, and hypernyms. For each new node, we add hypernym-hyponym edges between this new node and the matching previously added nodes while also removing any edges creating cycles in the domain graph. To avoid over-generalizing candidate labels, we perform only one or two iterations of the graph growing.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Automated term grouping and labeling</head><p>The goal of ANEA is to obtain (named) entity categories, i.e., few clusters of generally related terms of high cross-similarity and assign descriptive labels to these clusters. To do so, we maximize two parameters at the same time: cross-term group similarity and similarity between a group of terms and a label.</p><p>ANEA consists of the initial setup of the categories and three subsequent optimization steps to improve the representativeness of the terms and the assigned labels in the groups of terms.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.1">Setup.</head><p>We initialize ANEA by collecting all candidate categories, i.e., groups of potentially related terms and the assigned labels to them. Figure <ref type="figure" target="#fig_2">3</ref> depicts the process of candidate label collection. First, we iterate over all term-nodes, i.e., the domain graph's leaves and nodes that were created from the extracted terms not the hypernym WPs. For each term-node, we collect candidate labels extracted from the names of their hypernym-nodes. Each term obtains a list of candidate labels with various distances, i.e., the number of edges between a term-node and a label-node. We recursively traverse the domain graph as long as the distance between a term-node and a label-node is 𝑑 ≤ 𝑑 𝑚𝑎𝑥 , 𝑑 𝑚𝑎𝑥 = 5, or there are hypernyms to the current node in the domain graph. During the experiments, we notice that label-nodes with larger distances are often rather abstract and do not characterize a term well. Second, we "transpose" all terms and their candidate labels to obtain one label assigned to a group of terms. That is, we create a collection of categories among which we seek to find the most representative categories of the analyzed domain-specific text collection.</p><p>The selection of the most optimal categories among the candidates is a double-optimization process towards two requirements: generalization and specification. On the one hand, generalization aims at covering categories' broader semantics, e.g., a category with a more general label "Person" is better than categories such as "Actor, " "Politician, " etc. On the other hand, specification aims at selecting the category with more narrow semantics, e.g., categories such as "Country," "City," "State" provide more details about its terms than a category "Location".</p><p>We use a quality score 𝑄 𝑖 to evaluate each (entity) category 𝐸𝐶 𝑖 in a list of candidates:</p><formula xml:id="formula_0">𝑄 𝑖 = 𝑇 𝑖 • 𝐿 𝑖 • 𝑂 𝑖 • max(log 2 |𝐸𝐶 𝑖 |, 1) • 𝑑 𝑎𝑣𝑔_𝑖</formula><p>where 𝑇 𝑖 is a mean cross-term cosine similarity; 𝐿 𝑖 is a mean labelterms similarity; 𝑂 𝑖 is the overall similarity, i.e., 𝑂 𝑖 = 𝑇 𝑖 + 𝐿 𝑖 ; |𝐸𝐶 𝑖 | is a size of a category, i.e., the number of the terms in the class; 𝑑 𝑎𝑣𝑔_𝑖 is an average of non-zero distances between category's terms and label 𝑙:</p><formula xml:id="formula_1">𝑑 𝑎𝑣𝑔_𝑖 = 1 |𝐷 𝑛𝑛_𝑖 | 𝑑 ∈𝐷 𝑛𝑛_𝑖 𝑑, where 𝐷 𝑛𝑛_𝑖 = {𝐷 𝑖,𝑙 | ∀𝑖 ∈ 𝐸𝐶 : 𝐷 𝑖,𝑙 &gt; 0} and 𝐷 is a distance matrix 3 . If |𝐷 𝑛𝑛_𝑖 | = 0, then 𝑑 𝑎𝑣𝑔_𝑖 = 1.</formula><p>To calculate cosine similarities, we represent each term and label in a vector space with fastText word embeddings <ref type="bibr" target="#b7">[8]</ref>. We chose fastText for the representation of the out-of-vocabulary words, which often occur in domain-specific texts.</p><p>Requiring a large mean cross-term cosine similarity 𝑇 𝑖 increases the specificity of a category. Typically, the smaller the number of related terms, the larger the mean cross-similarity is. A larger mean label-terms similarity 𝐿 𝑖 also increases the specificity, i.e., a large similarity value is equivalent to a narrow descriptiveness of the terms of a label.</p><p>The overall sum 𝑂 𝑖 facilitates balancing potentially small values of either 𝑇 𝑖 or 𝐿 𝑖 if the other item is still large. The large size of a category increases its generalizing and descriptive properties, i.e., one label needs to describe as many terms as possible. Lastly, the average distance 𝑑 𝑎𝑣𝑔 _𝑖 acts as an amplifying factor for the generalization: the higher the label 𝑙 is in the domain graph, the more general is its meaning to the terms in this category.</p><p>Before the optimization steps, we perform filtering of the candidate categories to remove low quality categories from the candidate list. We remove an 𝐸𝐶 𝑖 if: (1) 𝑇 𝑖 &lt; 0.2, (2) 𝐿 𝑖 &lt; 0.3, (3)</p><formula xml:id="formula_2">|𝐸𝐶 𝑖 | &gt; 0.15 • |𝑇𝑇 𝐴| ∧ |𝐸𝐶 𝑖 | &lt; 5</formula><p>, where |𝑇𝑇 𝐴| is a number of termsto-annotate, i.e., a number of the term-nodes in the domain graph. In other words, we remove too vaguely related, very large or small categories. <ref type="figure" target="#fig_3">4a</ref> depicts that if two categories have the same terms but different labels, we sort the classes by their quality scores 𝑄 𝑖 and keep the categories with the highest 𝑄 𝑖 .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.2">Resolution of full overlaps. Figure</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>3.2.3</head><p>Resolution of substantial overlaps. Typically, categories have overlaps between their terms albeit we find that cross-terms and terms-label combinations of a single category are more semantically coherent than combinations of another category. We define that categories have substantial overlap if they share more than 50% of their terms.</p><p>Figure <ref type="figure" target="#fig_3">4b</ref> depicts the process of conflict resolution. We construct a matrix of replacements 𝑅, i.e., a matrix indicating the quality of a category measured by 𝑄 compared to those categories with substantial overlaps (values of 𝑅 initialized with 0). The matrix is used to identify if an 𝐸𝐶 𝐴 contains the best terms-label combination or there is a better 𝐸𝐶 𝑟𝑒𝑝𝑙 to replace 𝐸𝐶 𝐴 . Since ANEA's goal is to annotate as many generally related terms as possible yet find as specific categories as possible, we challenge both the size of 𝐸𝐶 𝑎 and its descriptive properties.</p><p>First, we sort categories by their 𝑄 score in decreasing order. Second, we intersect all categories with each other. If |𝐸𝐶 𝑎 ∩𝐸𝐶 𝑏 | ≥ 0.5 • |𝐸𝐶 𝑎 |, then we consider the overlaps substantial and add the quality score 𝑄 𝑏 to the matrix of replacements as a 𝑅 𝑎,𝑏 value. Note that the matrix is squared but asymmetrical because we calculate 50% of 𝐸𝐶 𝑎 's size and not of a pairwise function of two categories, e.g., min(|𝐸𝐶 𝑎 |, |𝐸𝐶 𝑏 |).</p><p>Finally, for each 𝐸𝐶 𝑎 : ∀𝑎 ∈ 𝐴 represented by a row 𝑟 𝑎 in 𝑅, we select a replacement 𝐸𝐶 𝑟𝑒𝑝𝑙 :</p><formula xml:id="formula_3">𝐸𝐶 𝑟𝑒𝑝𝑙 (𝐸𝐶 𝑎 ) = {𝐸𝐶 𝑐 | ∃𝑐 ∈ 𝐴 : arg max 𝑟 𝑎 = 𝐸𝐶 𝑐 ∧ arg max 𝑟 𝑐 = 𝐸𝐶 𝑐 }</formula><p>That is, we call 𝐸𝐶 𝑐 a replacement to 𝐸𝐶 𝑎 if 𝐸𝐶 𝑐 is the best among all comparable categories to 𝐸𝐶 𝑎 and also 𝐸𝐶 𝑐 is the best among all categories compared to itself. Also, a category can be a replacement for itself. We keep only the unique categories that are the best replacements {𝐸𝐶 𝑟𝑒𝑝𝑙 (𝐸 𝑎 ) : ∀𝑎 ∈ 𝐴}.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>3.2.4</head><p>Resolution of conflicting terms. After resolution of substantially overlapping terms, some categories contain minor conflicting terms, i.e., that are present in more than one category (Figure <ref type="figure" target="#fig_3">4c</ref>).</p><p>To resolve conflicting terms, first, we create a list of "clean" categories, i.e., from each category we remove all conflicting terms and record categories' labels from which the conflicting terms were removed. Additionally, we resolve the terms of the categories that may be also labels of another category, e.g., such a term as "h" in Figure <ref type="figure" target="#fig_2">3</ref>, i.e., move these label-terms to the categories with corresponding labels. We keep all categories even if some categories may afterward have no terms, i.e., if all their terms conflicted with other categories.</p><p>Second, to resolve conflicting terms, we estimate the quality of all "clean" categories. We calculate a quality score 𝑄 for each category (see Section 3.2; if |𝐸𝐶 | = 0, then 𝑄 = 0) and sort categories by decreasing 𝑄. Sorting brings forward categories that are the most probable to become final categories.</p><p>Third, we resolve all conflicting terms by beginning with those that belong to the categories with the highest 𝑄. For each conflicting term 𝑡 𝑗 and all 𝐸𝐶 𝑖 from which the term originated, we calculate a similarity score 𝑆: 𝑆 (𝑡 𝑗 , 𝐸𝐶 𝑖 ) = 𝑇 + 𝐿, where 𝑇 is the mean cosine similarity between the vector representation of 𝑡 𝑗 and the remaining terms in a "clean" 𝐸𝐶 𝑖 ; 𝐿 is the cosine similarity between 𝑡 𝑗 and a label of 𝐸𝐶 𝑖 . Even if a "clean" 𝐸𝐶 𝑖 contains no terms, i.e., 𝑇 = 0, 𝐿 will always yield 𝑆 &gt; 0. We select the best category for a given term 𝑡 𝑗 as:</p><formula xml:id="formula_4">𝐸𝐶 𝑏𝑒𝑠𝑡 (𝑡 𝑗 ) = {𝐸𝐶 𝑖 |∃𝑖 ∈ 𝐴 : arg max 𝑆 𝑖 (𝑡 𝑗 )}</formula><p>We add the resolved terms to their best matching category. The final categories are where |𝐸𝐶 | ≥ 5 , i.e., that represent a sufficiently large number of extracted terms from the given domain-specific texts.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">EXPERIMENTS</head><p>The evaluation goals are twofold. First, we seek to quantitatively assess the quality of the automatically extracted and annotated terms of both ANEA and a baseline using ratings from domain experts. Second, we seek to identify a recommendation for ANEA's default configuration to automatically annotate texts also of other domains by evaluating the annotated and human-assessed annotated datasets against silver quality datasets.</p><p>Due to the lack of German datasets for the analysis of domainspecific NER, we assess the quality of the produced categories through user studies where we ask users to rate the quality of the entities extracted by our system. We test ANEA and compare it to a baseline on four text datasets with four configurations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">User study</head><p>Our user study aimed at human assessment of the semantic quality of the produced categories due to different configurations. We collect feedback from human assessors for multiple configurations of two methods: ANEA and hierarchical clustering (see Section 4.2.1). First, we use this feedback to automatically construct silver-quality datasets and evaluate the proposed input configurations against them. Second, we use these silver datasets and evaluate the obtained configurations of the dataset for ANEA to find parameters for default configuration with which ANEA could be used to annotate other domain datasets.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.1">Test datasets.</head><p>We create four text datasets of comparable size from three different domains: processing industry (P), computer science, and traveling (T). To enable both cross-and intra-domain evaluation, we create two text datasets related to the computer science domain: databases (D) and software development (S). Table <ref type="table">1</ref> provides an overview of the datasets' parameters, such as the overall number of words, the number of unique terms and heads of terms (see Sec. 3.1), and the number of human assessors per each dataset. The table shows that the number of the unique heads may vary even given the identical number of unique extracted terms (cf. datasets S and T ). To test the applicability of the approach on different domains, we use the publicly available data from Wikipedia and a dataset built on private text data from a real-world production line in the processing industry. Specifically, the first three datasets (databases, software development, and traveling) originate from German Wikipedia articles dedicated to the respective categories. For each dataset, we searched for related articles in Wikipedia using a query "incategory:category", where "category" is "Datenbanken, " "Programmierung, " or "Reise". We iterated over the list of the search results sorted by relevance and extracted the texts of the articles if the articles had a specific number of words 𝑊 : 220 ≤ 𝑊 ≤ 2500, i.e., articles of medium size. The last dataset consists of reports about the daily operations of a company in the processing industry. Such reports include texts about statuses of the machinery, processes in the production lines, and problems that occurred throughout the daily routines. The dataset consists of approximately 200 short texts, each of 20-100 words.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.2">Experiment setup.</head><p>For the human assessment, we recruited nine native-German speaking participants (4 f, 5 m, aged between 23-60). Each participant is familiar with the domain of the assigned dataset(s) through their job, education, and/or hobbies.</p><p>We assigned 3-4 participants to each dataset, and each participant evaluated one or two datasets. Albeit the processing industry dataset has the smallest number of unique terms, we assigned the largest number of assessors to it due to the high relevance of obtaining valid results for such complex, expert domains as chemistry and technology. The vocabulary of these domains is typically strongly underrepresented in general text corpora used to train word embedding models <ref type="bibr" target="#b7">[8]</ref>.</p><p>The evaluation included two tasks for the participants: (1) assess the cross-term relatedness within the identified groups of terms and (2) assess the relatedness of the labels automatically assigned to the identified groups of terms. Per dataset, each participant needed to perform an assessment of eight sheets with automated annotation results: four identical input configurations per both ANEA and a baseline. Each participant needed to assign a semantic relatedness score between 0 and 9, where 0 meant no similarity and 9 -the highest similarity.</p><p>The input configuration included four different numbers of the input terms, i.e., terms-to-annotate (TTAs), among which the algorithms needed to extract the most representative terms that can form a separate semantic concept, i.e., an category. To vary the size of TTAs, we selected 1/𝑍 • 100% most frequent heads of phrases: 𝑍 ∈ [2, 3, 4, 5] for the datasets with a number of unique terms &lt; 1000, else 𝑍 ∈ <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b3">4,</ref><ref type="bibr" target="#b4">5,</ref><ref type="bibr" target="#b6">7]</ref>. By selecting only terms that share all most frequent heads, we ensure that these terms are the most representative of each domain-specific text.</p><p>Table <ref type="table">1</ref> reports the results of the user studies and shows that the relatedness of the groups of terms is higher than assigned labels. That is, the cross-term relatedness score was biased towards higher values, the label relatedness had a more uniform score distribution. Additionally, the mean and maximum of the relatedness scores vary across the datasets. We noticed that the relatedness scores were biased toward the size of the identified categories, i.e., a smaller number of terms in categories tend to have higher scores since it is easier for a human to assess a smaller number of items. However, we did not find any correlation between individual datasets and any of the outlined numeric characteristics.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>5: Construction of a silver dataset through collecting of information from user studies.</head><p>To estimate which input configuration and approach yielded the most coherent categories, we require a silver dataset, which will average the assigned scores and extract the highly rated combinations of terms into categories, and assigned the highly rated labels to them. 4.1.3 "Silver" datasets. The goal of a silver dataset is to ensure a fair and unified evaluation strategy of the approaches for all topics. We constructed a silver dataset for each topic by aggregating information from the human assessment sheets following the identical procedure.</p><p>First, for each dataset, we constructed term-to-term and label-toterms score matrices between the vocabulary of each topic and the extracted and assigned labels (Figure <ref type="figure">5</ref>). The matrices were initialized with zeros. We iterated over the relatedness scores across two approaches, four input configurations, and two-four human assessors. For every two terms in a term group, we added an assigned cross-term relatedness score to a value in a term-to-term matrix. This score demonstrates how two terms are evaluated in various combinations with other terms across different setups. After the summation was completed, we normalized each value in the matrix by the number of times two terms occurred together. We performed a similar procedure with the label-to-term relatedness scores: each term in a category and an assigned label to a category, we added a label-to-terms relatedness score assigned to a category and then normalized by the number of times a label was applied to a term.</p><p>To identify a threshold of relatedness of two terms belonging to a category, we built a histogram of all scores used to evaluate each dataset (see Table <ref type="table">1</ref>). In a score range of 0-9, we decided that a threshold of sufficient relevance of terms needs to lie higher than the mean score and not equal to the maximum value, i.e., between scores 6-8. Thus, for each dataset, we chose the most frequent score as a threshold and if a preceding score was less frequent by 1, then we calculated a mean of these scores.</p><p>We collected silver groups of terms, by choosing a term and merging it with other terms, a normalized relatedness score to which is higher or equal to the threshold of this dataset. We expanded this list of terms with the terms that are related to any of the merged terms compared to the threshold. Note that relatedness of all terms to each other exceeded a relatedness threshold, but relatedness of at least two terms needed to exceed a threshold. If a group of terms contained at least five terms, we form a silver category. We assigned a label to a silver category by ( <ref type="formula">1</ref>) calculating mean label-to-terms scores of all labels applied to at least two identified terms of a group, (2) selecting a label with the maximum mean score.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Evaluation</head><p>To evaluate the coherence and semantic quality of the produced categories at various input configurations, we introduce the evaluation methodology to evaluate ANEA and a representative baseline against the silver datasets. By identification the input configurations that yielded the best results, we sought to propose an optimal default ANEA's input configuration for any dataset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>4.2.1</head><p>Baseline: hierarchical clustering. We selected hierarchical clustering (HC) <ref type="bibr" target="#b14">[15]</ref> as a baseline to ANEA, since it successfully identifies semantically related terms that refer to identical entities <ref type="bibr" target="#b2">[3]</ref>. Although HC does not have the functionality of automated extraction and assignment of a label of a cluster of terms, we observed that HC's clusters could form meaningful categories. Therefore, we selected HC as a baseline to compare the quality of the produced groups of terms.</p><p>To ensure the best performance of HC per each dataset and each input configuration, we created an optimization of HC that selects the best clustering results and outputs clusters that contain maximum terms with maximum cross-term similarity. For each group of terms, we ran HC four times with fixed parameters of cosine similarity and average linkage criterion. We chose the linkage method, distance metric, and the optimization of the hyperparameters for the HC that are the most similar to the ANEA.</p><p>We built clustering configurations by varying the similarity threshold value between [0.5; 0.8] with a step of 0.1. For each clustering configuration 𝑗, we selected only clusters 𝐶𝐿 𝑗,𝑖 with more than 5 terms in each (|𝐶𝐿 𝑗,𝑖 | ≥ 5), i.e., impose the same minimum size requirements as for ANEA. Then, we calculated a weighted similarity score of each parameter configuration 𝑊 𝑆 𝑗 :</p><formula xml:id="formula_5">𝑊 𝑆 𝑗 = 1 𝐼 𝑖=0 |𝐶𝐿 𝑗,𝑖 | 𝐼 𝑖=0 𝑇 𝑗,𝑖 • |𝐶𝐿 𝑗,𝑖 |</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2021">-Workshop on Extraction and Evaluation Knowledge Entities from Scientific Documents</head><p>where 𝐼 is the number of clusters larger than 5 produced at a run 𝑗, and 𝑇 𝑗,𝑖 is a cross-term similarity within a cluster. We selected the best configuration as 𝐶 𝑏𝑒𝑠𝑡 = arg max 𝑗 (𝑊 𝑆 𝑗 • 𝐼 𝑖=0 |𝐶𝐿 𝑗,𝑖 |), i.e., a configuration that clusters the most terms and in the most semantically coherent way.</p><p>Since HC does not have label extraction and assigning functionality, the human assessors received only one task of assessment of the cross-term relatedness of clusters produced by HC.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.2">Metrics.</head><p>To evaluate the quality of the identified categories, we use five parameters: (1) number of categories: a larger number indicates diverse and narrowly defined categories, a smaller number -generalizing categories, (2) number of annotated terms (AT): property of identified relations between more terms, (3) the average size of categories: a smaller size indicates more narrowly-defined categories, whereas a larger size indicates more generally-related terms in categories; (4) average cross-term score (TS), ( <ref type="formula">5</ref>) average label-to-terms score (LS), and (6) average score (AS) between TS and LS: the high scores indicate the higher relatedness of the extracted terms and extracted and assigned labels. The main goal of our evaluation is to identify which input configurations lead to the highest average score between cross-term and label-to-terms relatedness while annotating more TTA into more general categories.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.3">Results.</head><p>To calculate average relatedness scores, we assigned the scores from the normalized score matrices to the terms and labels of the identified EC and average these scores. Table <ref type="table">2</ref> reports the evaluation results for four datasets and four input configurations. The table shows that HC gets the highest average relatedness score (𝐴𝑆 𝑎𝑣𝑔,𝐻𝐶 = 6.5) that almost reaches silver dataset (𝐴𝑆 𝑎𝑣𝑔,𝑠𝑖𝑙 𝑣𝑒𝑟 = 6.7) but at the same time produces categories of the size smaller than silver categories (𝑆𝑖𝑧𝑒 𝑎𝑣𝑔,𝐻𝐶 = 6 and 𝑆𝑖𝑧𝑒 𝑎𝑣𝑔,𝑠𝑖𝑙 𝑣𝑒𝑟 = 16). While on average, ANEA annotates the largest number of terms (𝐴𝑇 𝑚𝑒𝑎𝑛,𝐴𝑁 𝐸𝐴 = 175), it also yields the lowest average relatedness score (𝐴𝑆 𝑚𝑒𝑎𝑛,𝐴𝑁 𝐸𝐴 = 5.2) both compared to the silver dataset and HC. When creating a coding book, multiple human coders first annotate, and then the majority voting decides which excepts and labels describe a dataset the best. We applied a similar strategy to improve the performance of ANEA.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>4.2.4</head><p>Voting strategy and default input configuration. Ensemble learning is a common approach in machine learning to improve the results of a classifier, i.e., by combining predictions of multiple classifiers to achieve a boost in the overall accuracy through collecting "the wisdom of the crowd" <ref type="bibr" target="#b1">[2]</ref>.</p><p>We followed this principle to improve the quality of the extracted by ANEA categories and combine results of 2-4 configurations in each dataset. Similar to the construction of silver datasets, we created a category of at least five terms if these terms co-occurred in at least two input configurations. We assigned a label that describes the majority of the terms in the identified group.</p><p>Table <ref type="table">3</ref> reports that at least one of the combination of ANEA with multiple input configurations increases the average relatedness score compared to ANEA without a voting strategy on average by 0.7 (𝐴𝑆 𝑎𝑣𝑔,𝐴𝑁 𝐸𝐴 𝑣𝑜𝑡𝑒 = 5.9). Although the voting approach does not exceed the relatedness scores of the silver datasets (reaches 87.9% of the silver score), it increases a number of annotated terms per category (𝐴𝑆 𝑎𝑣𝑔,𝐴𝑁 𝐸𝐴 = 11 and 𝐴𝑆 𝑎𝑣𝑔,𝐴𝑁 𝐸𝐴 𝑣𝑜𝑡𝑒 = 13) and Table <ref type="table">2</ref>: Evaluation of the entity identification methods in various input configuration. Z is the denominator to choose terms-to-annotate from the most 1/𝑍 frequent terms' heads; TTA is the number of terms-to-annotate, EC is the number of the (entity) categories that were produced by each approach; AT is the number of annotated terms, i.e., that belong to the identified categories; TS is the mean cross-term relatedness score among the categories; LS is the mean label-to-terms relatedness score; AS is the average score between TS and LS (* means that AS is equal to TS because HC does not assign labels to groups of terms To identify the best default configuration for the voting strategy of ANEA, we selected the best performing voting strategy configurations per each dataset and deduced a default input configuration by generalizing these configurations. We took the minimum and Table <ref type="table">3</ref>: Evaluation of the entity identification methods when (entity) categories (EC) are created by majority voting between input configurations. In each dataset, the majority voting improves relatedness score for each dataset. The highlighted configurations show a range of terms-to-annotate (TTA) and the resulted best average score (AS).   <ref type="table">1</ref>). Figure <ref type="figure" target="#fig_6">6</ref> depicts a linear trend between the TTAs and the unique heads from the datasets. Based on this trend, for any other dataset, we recommend annotating only the first 𝑦 = 158 + 0.167𝑥 TTAs that belong to the most frequent heads, where 𝑥 is a number of unique heads in a dataset. For the voting strategy, we recommend using the input configurations of the 𝑦, 𝑦 − 40, and 𝑦 + 40 number of terms that share the most frequent heads of terms.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">DISCUSSION AND FUTURE WORK</head><p>Our evaluation shows that ANEA facilitates a faster annotation process. Specifically, ANEA automatically performs the most timeconsuming tasks of deriving a coding book for the annotation of a dataset for NER. ANEA imitates the first two stages of a manual annotation process. First, a small set of articles (≈6000-8000 words) is used to automatically identify categories relevant to the data of the current domain, i.e., identify and extract related terms and assign a label to each of them (identified 4-12 categories with 44-157 assigned terms). Second, a voting strategy is applied, which aims at increasing the validity of the derived categories by following the idea of ensemble learning and intercoder agreement (relatedness score improved on average from 77.4% of the silver average relatedness score to 87.9%). To continue with manual annotation of a (N)ER dataset, next, researchers manually validate their coding book. If they find that the coding book sufficiently represents the dataset, they annotate the remaining texts to create a large corpus for NER. Therefore, the primary use cases of ANEA are as follows. First, extraction of domain categories from a subset of a large text dataset and improve their quality with the voting strategy. Second, manual validation and improvement of the identified categories by moving terms between the categories and suggesting better labels to them.</p><p>The final stage of annotation of a NER dataset is to apply a coding book to a large dataset, i.e., read the text and assign categories to text excerpts following guidelines or examples of the coding book. Although, such manual text annotation is a standard approach to create "gold"-standard datasets, the recent semi-supervised learning neural network models (e.g., DART <ref type="bibr" target="#b3">[4]</ref>) show high potential in reliable annotating of large text collections. We plan to use the models like DART will complete the automation of creating domainspecific NER datasets.</p><p>Future work directions include the creation of manually annotated datasets from scratch for multiple domains to calculate accuracy metrics, e.g., precision, recall, and F1, to evaluate the effectiveness of identifying categories by ANEA. Further, to improve the quality and meaningfulness of the assigned labels, we plan to test ANEA on other knowledge graphs, e.g., Wikidata or BabelNet. To test the applicability of ANEA, we also plan to evaluate the approach in other languages, e.g., English, with an additional module for the identification of multi-word expressions similar to compound-based German words <ref type="bibr" target="#b19">[20]</ref>. Further, to improve the semantic quality of 2021 -Workshop on Extraction and Evaluation of Knowledge Entities from Scientific Documents both categories' terms and labels in a specific domain, we plan to use a language model, e.g., BERT, use the quality score as a learning objective. Lastly, we seek to build a semi-supervised NER model to complete automated annotation of NER datasets, i.e., automatically annotate large datasets suitable for training neural network models. We will use terms and labels from the derived categories as the seed-terms and seed-labels and perform named entity tagging and classify more domain terms <ref type="bibr" target="#b3">[4,</ref><ref type="bibr" target="#b22">23]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">CONCLUSION</head><p>In this paper, we propose ANEA, an automatic approach to derive domain entity categories from a subset of domain input texts, i.e., create a small dataset to train a NER model. Specifically, ANEA identifies related domain-representative terms and automatically extracts and assigns descriptive and generalizing labels to them based on Wiktionary. In our user assessment and evaluation, ANEA could not outperform a silver dataset on the relatedness scores assigned to the groups terms and labels describing these groups. However, ANEA produced more generalizing domain categories compared to a strong baseline. We showed that our voting strategy of combining terms and labels from the categories identified at multiple input configurations significantly improved the quality of the final categories. Additionally, we suggested a default input configuration that can be applied to derive categories from German domain text datasets. Finally, we think that the best application of ANEA is to annotate and use a small dataset in semi-supervised learning. Moreover, we plan to improve and validate the annotations with a domain expert, and use this small domain dataset to train state-of-the-art NER models.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: An example of a Wiktionary page (WP).</figDesc><graphic coords="2,60.43,479.74,226.98,158.15" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: (a) A domain graph is a combination of Wiktionary and domain terminology extracted from text. (b) Domain graph construction: initialization and expansion of the graph.</figDesc><graphic coords="3,66.41,83.68,479.18,245.47" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: ANEA setup: a collection of candidate entity categories.</figDesc><graphic coords="4,73.04,421.37,201.76,217.53" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: (a) Resolution of the full overlaps. (b) Resolution of the substantial overlaps. (c) Resolution of the conflicting terms.</figDesc><graphic coords="4,324.59,83.69,226.98,334.57" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Table 1 :</head><label>1</label><figDesc>Dataset statistics: databases (D), software development (S), travelling (T), and processing industry (P). The histograms show the distribution of the user-assigned relatedness score to categories extracted with various input configurations per dataset. Bold scores indicate a threshold used to construct silver datasets.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head>Figure 6 :</head><label>6</label><figDesc>Figure 6: Default input configuration: a number of terms-to-annotate linearly depends on the number of unique NP heads per dataset.</figDesc><graphic coords="9,324.59,83.69,226.98,137.73" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0"><head></head><label></label><figDesc></figDesc><graphic coords="7,66.41,83.69,479.20,146.18" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head></head><label></label><figDesc>).</figDesc><table><row><cell>topic Appr.</cell><cell cols="6">Z TTA EC AT Size TS LS AS</cell></row><row><cell>silver</cell><cell>3 420</cell><cell>5</cell><cell cols="2">113 23</cell><cell cols="2">7.2 7.0 7.2</cell></row><row><cell></cell><cell>3 420</cell><cell cols="3">14 108 8</cell><cell>6.9 -</cell><cell>6.9*</cell></row><row><cell>HC</cell><cell>4 363 5 316</cell><cell cols="2">12 87 10 73</cell><cell>7 7</cell><cell>7.0 -7.0 -</cell><cell>7.0* 7.0*</cell></row><row><cell>D</cell><cell>7 253</cell><cell>8</cell><cell>52</cell><cell>7</cell><cell>7.2 -</cell><cell>7.2*</cell></row><row><cell></cell><cell>3 420</cell><cell cols="3">26 306 12</cell><cell cols="2">4.7 4.2 4.4</cell></row><row><cell>ANEA</cell><cell>4 363 5 316</cell><cell cols="3">22 255 12 21 234 11</cell><cell cols="2">5.3 4.6 5.0 5.6 5.0 5.3</cell></row><row><cell></cell><cell>7 253</cell><cell cols="3">18 179 10</cell><cell cols="2">5.7 5.0 5.4</cell></row><row><cell>silver</cell><cell>3 356</cell><cell>6</cell><cell>57</cell><cell>10</cell><cell cols="2">6.2 6.0 6.1</cell></row><row><cell></cell><cell>3 356</cell><cell cols="3">17 190 11</cell><cell>5.3 -</cell><cell>5.3*</cell></row><row><cell>HC</cell><cell>4 303 5 255</cell><cell cols="3">15 152 10 15 137 9</cell><cell>5.5 -5.4 -</cell><cell>5.5* 5.4*</cell></row><row><cell>S</cell><cell>7 191</cell><cell cols="3">12 103 9</cell><cell>5.4 -</cell><cell>5.4*</cell></row><row><cell></cell><cell>3 356</cell><cell cols="3">21 242 12</cell><cell cols="2">4.1 4.4 4.2</cell></row><row><cell>ANEA</cell><cell>4 303 5 255</cell><cell cols="3">18 186 10 15 164 11</cell><cell cols="2">4.2 3.8 4.0 4.2 4.4 4.3</cell></row><row><cell></cell><cell>7 191</cell><cell cols="3">10 119 12</cell><cell cols="2">5.0 5.3 5.2</cell></row><row><cell>silver</cell><cell>3 363</cell><cell>6</cell><cell cols="2">115 19</cell><cell cols="2">7.8 6.7 7.3</cell></row><row><cell></cell><cell>3 363</cell><cell cols="3">19 156 8</cell><cell>7.3 -</cell><cell>7.3*</cell></row><row><cell>HC</cell><cell>4 297 5 258</cell><cell cols="3">16 133 8 12 105 9</cell><cell>7.3 -7.3 -</cell><cell>7.3* 7.3*</cell></row><row><cell>T</cell><cell>7 211</cell><cell>9</cell><cell>76</cell><cell>8</cell><cell>7.2 -</cell><cell>7.2*</cell></row><row><cell></cell><cell>3 363</cell><cell cols="3">22 239 11</cell><cell cols="2">5.4 4.8 5.1</cell></row><row><cell>ANEA</cell><cell>4 297 5 258</cell><cell cols="3">17 191 11 14 161 12</cell><cell cols="2">5.4 4.4 4.9 5.3 4.4 4.9</cell></row><row><cell></cell><cell>7 211</cell><cell cols="3">14 137 10</cell><cell cols="2">5.6 4.0 4.8</cell></row><row><cell>silver</cell><cell>2 282</cell><cell>7</cell><cell cols="2">102 15</cell><cell cols="2">6.6 6.2 6.4</cell></row><row><cell></cell><cell>2 282</cell><cell>9</cell><cell>65</cell><cell>7</cell><cell>5.2 -</cell><cell>5.2*</cell></row><row><cell>HC</cell><cell>3 227 4 200</cell><cell>8 8</cell><cell>61 61</cell><cell>8 8</cell><cell>6.0 -6.0 -</cell><cell>6.0* 6.0*</cell></row><row><cell>P</cell><cell>5 183</cell><cell>7</cell><cell>56</cell><cell>8</cell><cell>6.1 -</cell><cell>6.1*</cell></row><row><cell></cell><cell>2 282</cell><cell cols="3">18 213 12</cell><cell cols="2">4.7 4.5 4.6</cell></row><row><cell>ANEA</cell><cell>3 227 4 200</cell><cell cols="3">16 172 11 16 163 10</cell><cell cols="2">5.3 4.9 5.1 5.3 4.8 5.1</cell></row><row><cell></cell><cell>5 183</cell><cell cols="3">15 149 10</cell><cell cols="2">5.2 4.5 4.8</cell></row><row><cell cols="7">also identifies more generalizing categories (𝐸𝐶 𝑎𝑣𝑔,𝐴𝑁 𝐸𝐴 = 15 and</cell></row><row><cell cols="2">𝐸𝐶 𝑎𝑣𝑔,𝐴𝑁 𝐸𝐴 𝑣𝑜𝑡𝑒 = 9).</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_0">We extract the tokens that have one of the following dependency tags: "ROOT", "oa"= accusative object, "oa2" = second accusative object, "app" = apposition, "cj" = conjunct.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_1">Rows are the names of term-nodes and the columns are the names of term-nodes and label-nodes. The columns contain also the term-nodes because some of the term-nodes may not be leaves but nodes of the domain graph (see Figure2).</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>ACKNOWLEDGEMENT</head><p>The research for this paper has been conducted in collaboration with the company eschbach (https://eschbach.com) supported by the Central Innovation Programme (ZIM) of the German Federal Ministry for Economic Affairs and Energy.</p><p>We thank all study participants for their significant contribution to this publication.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">NoSta-D Named Entity Annotation for German: Guidelines and Dataset</title>
		<author>
			<persName><forename type="first">Darina</forename><surname>Benikova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Chris</forename><surname>Biemann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Marc</forename><surname>Reznicek</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">LREC</title>
				<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="2524" to="2531" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title level="m" type="main">Guide to intelligent data analysis: how to intelligently make sense of real data</title>
		<author>
			<persName><forename type="first">Christian</forename><surname>Michael R Berthold</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Frank</forename><surname>Borgelt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Frank</forename><surname>Höppner</surname></persName>
		</author>
		<author>
			<persName><surname>Klawonn</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2010">2010</date>
			<publisher>Springer Science &amp; Business Media</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">SenticNet 4: A semantic resource for sentiment analysis based on conceptual primitives</title>
		<author>
			<persName><forename type="first">Erik</forename><surname>Cambria</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Soujanya</forename><surname>Poria</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Rajiv</forename><surname>Bajpai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Björn</forename><surname>Schuller</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of COLING 2016, the 26th international conference on computational linguistics: Technical papers</title>
				<meeting>COLING 2016, the 26th international conference on computational linguistics: Technical papers</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="2666" to="2677" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">DART: A Lightweight Quality-Suggestive Data-to-Text Annotation Tool</title>
		<author>
			<persName><forename type="first">Ernie</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jeriah</forename><surname>Caplinger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Alex</forename><surname>Marin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Xiaoyu</forename><surname>Shen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Vera</forename><surname>Demberg</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2020.coling-demos.3</idno>
		<ptr target="https://doi.org/10.18653/v1/2020.coling-demos.3" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 28th International Conference on Computational Linguistics: System Demonstrations. International Committee on Computational Linguistics (ICCL)</title>
				<meeting>the 28th International Conference on Computational Linguistics: System Demonstrations. International Committee on Computational Linguistics (ICCL)<address><addrLine>Barcelona, Spain</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="12" to="17" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Chemical named entities recognition: a review on approaches and applications</title>
		<author>
			<persName><forename type="first">Safaa</forename><surname>Eltyeb</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Naomie</forename><surname>Salim</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of cheminformatics</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page">17</biblScope>
			<date type="published" when="2014">2014. 2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Unsupervised named-entity extraction from the web: An experimental study</title>
		<author>
			<persName><forename type="first">Oren</forename><surname>Etzioni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Michael</forename><surname>Cafarella</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Doug</forename><surname>Downey</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ana-Maria</forename><surname>Popescu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Tal</forename><surname>Shaked</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Stephen</forename><surname>Soderland</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Alexander</forename><surname>Daniel S Weld</surname></persName>
		</author>
		<author>
			<persName><surname>Yates</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Artificial intelligence</title>
		<imprint>
			<biblScope unit="volume">165</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="91" to="134" />
			<date type="published" when="2005">2005. 2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m" type="main">Named Entity Recognition with Extremely Limited Data</title>
		<author>
			<persName><forename type="first">John</forename><surname>Foley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Sheikh</forename><surname>Muhammad Sarwar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">James</forename><surname>Allan</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1806.04411</idno>
		<imprint>
			<date type="published" when="2018">2018. 2018</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Learning Word Vectors for 157 Languages</title>
		<author>
			<persName><forename type="first">Édouard</forename><surname>Grave</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Piotr</forename><surname>Bojanowski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Prakhar</forename><surname>Gupta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Armand</forename><surname>Joulin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Tomáš</forename><surname>Mikolov</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC</title>
				<meeting>the Eleventh International Conference on Language Resources and Evaluation (LREC</meeting>
		<imprint>
			<date type="published" when="2018">2018. 2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">A comparison of NER tools wrt a domain-specific vocabulary</title>
		<author>
			<persName><forename type="first">Timm</forename><surname>Heuss</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Bernhard</forename><surname>Humm</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Christian</forename><surname>Henninger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Thomas</forename><surname>Rippl</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 10th International Conference on Semantic Systems</title>
				<meeting>the 10th International Conference on Semantic Systems</meeting>
		<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="100" to="107" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">spaCy 2: Natural language understanding with Bloom embeddings, convolutional neural networks and incremental parsing</title>
		<author>
			<persName><forename type="first">Matthew</forename><surname>Honnibal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ines</forename><surname>Montani</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2017">2017. 2017</date>
		</imprint>
	</monogr>
	<note>To appear</note>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Curex: a system for extracting, curating, and exploring domain-specific knowledge graphs from text</title>
		<author>
			<persName><forename type="first">Michael</forename><surname>Loster</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Felix</forename><surname>Naumann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jan</forename><surname>Ehmueller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Benjamin</forename><surname>Feldmann</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 27th ACM International Conference on Information and Knowledge Management</title>
				<meeting>the 27th ACM International Conference on Information and Knowledge Management</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="1883" to="1886" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Improving Company Recognition from Unstructured Text by using Dictionaries</title>
		<author>
			<persName><forename type="first">Michael</forename><surname>Loster</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Zhe</forename><surname>Zuo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Felix</forename><surname>Naumann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Oliver</forename><surname>Maspfuhl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Dirk</forename><surname>Thomas</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">EDBT</title>
				<imprint>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="610" to="619" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Named entity recognition approaches</title>
		<author>
			<persName><forename type="first">Alireza</forename><surname>Mansouri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lilly</forename><surname>Suriani Affendey</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ali</forename><surname>Mamat</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Computer Science and Network Security</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="339" to="344" />
			<date type="published" when="2008">2008. 2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">What psycholinguists know about chemistry: Aligning Wiktionary and WordNet for increased domain coverage</title>
		<author>
			<persName><forename type="first">M</forename><surname>Christian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Iryna</forename><surname>Meyer</surname></persName>
		</author>
		<author>
			<persName><surname>Gurevych</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of 5th International Joint Conference on Natural Language Processing</title>
				<meeting>5th International Joint Conference on Natural Language Processing</meeting>
		<imprint>
			<date type="published" when="2011">2011</date>
			<biblScope unit="page" from="883" to="892" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Algorithms for hierarchical clustering: an overview</title>
		<author>
			<persName><forename type="first">Fionn</forename><surname>Murtagh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Pedro</forename><surname>Contreras</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="86" to="97" />
			<date type="published" when="2012">2012. 2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Unsupervised namedentity recognition: Generating gazetteers and resolving ambiguity</title>
		<author>
			<persName><forename type="first">David</forename><surname>Nadeau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Peter</forename><forename type="middle">D</forename><surname>Turney</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Stan</forename><surname>Matwin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Conference of the Canadian society for computational studies of intelligence</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="266" to="277" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Mining wiki resources for multilingual named entity recognition</title>
		<author>
			<persName><forename type="first">E</forename><surname>Alexander</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Patrick</forename><surname>Richman</surname></persName>
		</author>
		<author>
			<persName><surname>Schone</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of ACL-08: HLT</title>
				<meeting>ACL-08: HLT</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="1" to="9" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">ChemSpot: a hybrid system for chemical named entity recognition</title>
		<author>
			<persName><forename type="first">Tim</forename><surname>Rocktäschel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Michael</forename><surname>Weidlich</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ulf</forename><surname>Leser</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Bioinformatics</title>
		<imprint>
			<biblScope unit="volume">28</biblScope>
			<biblScope unit="page" from="1633" to="1640" />
			<date type="published" when="2012">2012. 2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<title level="m" type="main">Fine-grained Named Entity Annotations for German Biographic Interviews</title>
		<author>
			<persName><forename type="first">Josef</forename><surname>Ruppenhofer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ines</forename><surname>Rehbein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Carolina</forename><surname>Flinz</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2020">2020. 2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Detecting noncompositional mwe components using wiktionary</title>
		<author>
			<persName><forename type="first">Bahar</forename><surname>Salehi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Paul</forename><surname>Cook</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Timothy</forename><surname>Baldwin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP)</title>
				<meeting>the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP)</meeting>
		<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="1792" to="1797" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Introduction to the CoNLL-2003 Shared Task: Language-Independent Named Entity Recognition</title>
		<author>
			<persName><forename type="first">Erik</forename><surname>Tjong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kim</forename><surname>Sang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Fien</forename><surname>De Meulder</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Seventh Conference on Natural Language Learning at HLT-NAACL</title>
				<meeting>the Seventh Conference on Natural Language Learning at HLT-NAACL</meeting>
		<imprint>
			<date type="published" when="2003">2003. 2003</date>
			<biblScope unit="page" from="142" to="147" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">A German Corpus for Fine-Grained Named Entity Recognition and Relation Extraction of Traffic and Industry Events</title>
		<author>
			<persName><forename type="first">Martin</forename><surname>Schiersch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Veselina</forename><surname>Mironova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Maximilian</forename><surname>Schmitt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Philippe</forename><surname>Thomas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Aleksandra</forename><surname>Gabryszak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Leonhard</forename><surname>Hennig</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC</title>
				<meeting>the Eleventh International Conference on Language Resources and Evaluation (LREC</meeting>
		<imprint>
			<date type="published" when="2018">2018. 2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Learning Named Entity Tagger using Domain-Specific Dictionary</title>
		<author>
			<persName><forename type="first">Jingbo</forename><surname>Shang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Liyuan</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Xiaotao</forename><surname>Gu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Xiang</forename><surname>Ren</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Teng</forename><surname>Ren</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jiawei</forename><surname>Han</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing</title>
				<meeting>the 2018 Conference on Empirical Methods in Natural Language Processing</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="2054" to="2064" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Automatically generated NE tagged corpora for English and Hungarian</title>
		<author>
			<persName><forename type="first">Eszter</forename><surname>Simon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Dávid</forename><surname>Márk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Nemeskey</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Association for Computational Linguistics</title>
				<imprint>
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Domain specific named entity recognition referring to the real world by deep neural networks</title>
		<author>
			<persName><forename type="first">Suzushi</forename><surname>Tomori</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Takashi</forename><surname>Ninomiya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Shinsuke</forename><surname>Mori</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics</title>
		<title level="s">Short Papers</title>
		<meeting>the 54th Annual Meeting of the Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="236" to="242" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<monogr>
		<title level="m" type="main">Incremental Coreference Resolution for German</title>
		<author>
			<persName><forename type="first">Don</forename><surname>Tuggener</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
	<note type="report_type">Ph. D. Dissertation</note>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Web-based ontology learning with isolde</title>
		<author>
			<persName><forename type="first">Nicolas</forename><surname>Weber</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Paul</forename><surname>Buitelaar</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the ISWC Workshop on Web Content Mining with Human Language Technologies</title>
				<meeting>of the ISWC Workshop on Web Content Mining with Human Language Technologies</meeting>
		<imprint>
			<publisher>Citeseer</publisher>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">A Survey on Recent Advances in Named Entity Recognition from Deep Learning models</title>
		<author>
			<persName><forename type="first">Vikas</forename><surname>Yadav</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Steven</forename><surname>Bethard</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 27th International Conference on Computational Linguistics</title>
				<meeting>the 27th International Conference on Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="2145" to="2158" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">Extracting Lexical Semantic Knowledge from Wikipedia and Wiktionary</title>
		<author>
			<persName><forename type="first">Torsten</forename><surname>Zesch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Christof</forename><surname>Müller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Iryna</forename><surname>Gurevych</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">LREC</title>
				<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="page" from="1646" to="1652" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<analytic>
		<title level="a" type="main">Unsupervised biomedical named entity recognition: Experiments with clinical and biological texts</title>
		<author>
			<persName><forename type="first">Shaodian</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Noémie</forename><surname>Elhadad</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of biomedical informatics</title>
		<imprint>
			<biblScope unit="volume">46</biblScope>
			<biblScope unit="issue">6</biblScope>
			<biblScope unit="page" from="1088" to="1098" />
			<date type="published" when="2013">2013. 2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<analytic>
		<title level="a" type="main">Automatically building largescale named entity recognition corpora from Chinese Wikipedia</title>
		<author>
			<persName><forename type="first">Jie</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Bi-Cheng</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gang</forename><surname>Chen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Frontiers of Information Technology &amp; Electronic Engineering</title>
		<imprint>
			<biblScope unit="volume">16</biblScope>
			<biblScope unit="issue">11</biblScope>
			<biblScope unit="page" from="940" to="956" />
			<date type="published" when="2015">2015. 2015</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
