<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">VALIDATION OF MIXED-STRUCTURED DATA USING PATTERN MINING AND INFORMATION EXTRACTION</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Martin</forename><surname>Atzmueller</surname></persName>
							<email>atzmueller@cs.uni-kassel.de</email>
							<affiliation key="aff0">
								<orgName type="institution">University of Kassel Knowledge and Data Engineering</orgName>
								<address>
									<settlement>Kassel</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Stephanie</forename><surname>Beer</surname></persName>
							<email>beer_s@klinik.uni-wuerzburg.de</email>
							<affiliation key="aff1">
								<orgName type="institution">University-Hospital of Würzburg Gastroontologics Research Group</orgName>
								<address>
									<settlement>Würzburg</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">VALIDATION OF MIXED-STRUCTURED DATA USING PATTERN MINING AND INFORMATION EXTRACTION</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">33E52D1751A7DA074E9AA1EE7586DBD5</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T12:32+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>For large-scale data mining utilizing data from ubiquitous and mixed-structured data sources, the appropriate extraction and integration into a comprehensive data-warehouse is of prime importance. Then, appropriate methods for validation and potential refinement are essential. This paper presents an approach applying data mining and information extraction methods for data validation: We apply subgroup discovery and (rule-based) information extraction for data integration and validation. The methods are integrated into an incremental process for continuous validation options. The results of a medical application demonstrate that subgroup discovery and the applied information extraction methods are well suited for mining, extracting and validating clinically relevant knowledge.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">INTRODUCTION</head><p>Whenever data is continously collected, for example, using intelligent documentation systems <ref type="bibr" target="#b1">[1]</ref>, data mining and data analysis provide a broad range of options for scientific purposes. The mining and analysis step is often implemented using a data-warehouse <ref type="bibr" target="#b2">[2,</ref><ref type="bibr" target="#b3">3,</ref><ref type="bibr" target="#b4">4]</ref>. For the data preprocessing and integration of several heterogenous sources, there exist standardized extracttransform-load (ETL) procedures, that need to incorporate suitable data schemas, and integration rules. Additionally, for unstructured or semi-structured textual data sources, the integration requires effective information extraction methods. For clinical discharge letters, for example, the structure of the letter is usually nonstandardized, and thus dependent on different writing styles of different authors.</p><p>However, a prerequisite of data mining is the validation and the quality assurance of the integrated data. Especially concerning unreliable extraction and integration methods, the quality of the obtained data can vary significantly. If the data has been successfully validated, then the trust in the data mining results and their acceptance can be increased.</p><p>In this paper, we propose an approach for the validation of mixed-structured data using data mining and information extraction and propose appropriate refinement options. We focus on a data mining technique for mining local patterns, i.e., subgroup discovery, e.g., <ref type="bibr">[5,</ref><ref type="bibr" target="#b6">6,</ref><ref type="bibr" target="#b8">7]</ref> that are especially suitable for the task: Local patterns consider local regularities (and irregularities) of the data and are therefore useful for spotting nonexpected, contradicting, and otherwise unusual patterns potentially indicating problems and errors in the data.</p><p>Concerning the information extraction techniques, we consider popular methods implemented in the UIMA <ref type="bibr" target="#b9">[8]</ref> and ClearTK <ref type="bibr" target="#b10">[9]</ref> framework, and especially focus on the TEXTMARKER system, e.g., <ref type="bibr" target="#b11">[10,</ref><ref type="bibr" target="#b12">11]</ref> for rulebased information extraction. Rules are especially suitable for the proposed information extraction task since they allow a concise and declarative formalization of the relevant domain knowledge that is especially easy to acquire, to comprehend and to maintain. Furthermore, in the case of errors, the cause can easily be identified by tracing the application of the individual rules.</p><p>The combined approach enables data mining from heterogenous sources. The user can specify simple rules that consider features of the text, e.g., structural or syntactic features of the textual content. We focus on an incremental level-wise approach, such that both methods can complement each other in the validation and refinement setting. Furthermore, validation knowledge can be formalized in a knowledge base, for assessing known and expected relations in the data.</p><p>The approach has been implemented in a clinical application for mining data from clinical information systems, documentation systems, and clinical discharge letters. This application scenario concerns the data integration from heterogenous databases and the information extraction from textual documents. The experiences and results so far demonstrate the flexibility and effectiveness of the presented approach that make the data mining and information extraction methods suitable components in the mining, validation and refinement process.</p><p>In the following, we shortly summarize the methods for data mining and information extraction, subgroup discovery, and rule-based information extraction using TEXTMARKER.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Subgroup Discovery</head><p>Subgroup discovery is a flexible data mining method for discovering local patterns that can be utilized for global modeling in the context of exploratory data analysis, description, characterization and classification.</p><p>Subgroup discovery is applied for identifying relations between a (dependent) target concept and a set of explaining (independent) variables. Then, the goal is to describe subsets of the data, that have the most unusual characteristics with respect to the concept of interest given by the target variable <ref type="bibr" target="#b6">[6]</ref>. For example, the risk of coronary heart disease (target variable) is significantly higher in the subgroup of smokers with a positive family history than in the general population.</p><p>In the context of the proposed validation approach, we consider certain gold-standard concepts as targets, as well as target concepts that are true, if and only if equivalent concepts from two different sources match. Then, we can identify combinations of factors that cause a mismatch between the concepts. These combinations can then indicate candidates for refinement.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Rule-based Information Extraction</head><p>Information extractions aims at extracting a set of concepts, entities and relations from a set of documents. TEXTMARKER <ref type="bibr" target="#b11">[10,</ref><ref type="bibr" target="#b12">11]</ref> is a robust system for rulebased information extraction. It can be applied very intuitively, since the used rules are especially easy to acquire and to comprehend. Using the extracted information, data records can be easily created in a postprocessing step. Humans often apply a strategy according to a highlighter metaphor during 'manual' information extraction: First, top-level text blocks are considered and classified according to their content by coloring them with different highlighters. The contained elements of the annotated texts segments are then considered further. The TEXTMARKER <ref type="bibr" target="#b11">[10,</ref><ref type="bibr" target="#b12">11]</ref> system tries to imitate this manual extraction method by formalizing the appropriate actions using matching rules:</p><p>The rules mark sequences of words, extract text segments or modify the input document depending on textual features.</p><p>TextMarker aims at supporting the knowledge engineer in the rapid prototyping of information extraction applications. The default input for the system is semi-structured text, but it can also process structured or free text. Technically, HTML is often the input format, since most word processing documents can be obtained in HTML format, or converted appropriately.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">THE MINING AND VALIDATION PROCESS</head><p>Figure <ref type="figure" target="#fig_0">1</ref> depicts the process of validation and refinement of mixed-structured data using pattern mining and information extraction methods. The input of the process is given by data from heterogenous data sources, and by textual documents. The former are processed by appropriate data integration methods adapted to the different sources. The latter are handled by information extraction techniques, e.g., rule-based methods that utilize appropriate extraction rules for the extraction of concepts and relations from the documents. In general, a variety of methods can be applied.</p><p>The process supports arbitrary information extraction methods, e.g., automatic techniques like supportvector machines or conditional random fields as implemented in the ClearTK <ref type="bibr" target="#b10">[9]</ref> toolkit for statistical natural language processing. However, the refinement capabilies vary for the different extraction approaches: While black-box methods like support vector machines or conditional random fields only allow an indirect refinement and adaptation of the model, i.e., based on adapting the input data and/or the method parameters for constructing the model, a white-box approach implemented using rules provides for a direct modification of its model, namely the provided rules. Therefore, we especially focus on rule-based methods due to their rich refinement capabilities.</p><p>After the integration and extraction of the data, the result is provided to the pattern mining system which obtains a set of validation patterns as output. This set is then checked both for internal consistency and compared to formalized background knowledge. In the case of discrepancies and errors, refinement are proposed for the data integration and/or the information extraction steps. After the rules have been refined, the process iterates with the updated schemas and models.</p><p>In the following we discuss exemplary results obtained from a medical project. We applied data collected by the SONOCONSULT system, a multifunctional knowledge system for sonography, which has been in routine use since 2002 documenting more than 12000 patients in two clinics. The system covers the entire field of abdominal ultrasound (liver, portal tract, gallbladder, spleen, kidneys, adrenal glands, pancreas, intestine, lymph nodes, abdominal aorta, cava inferior, prostate, and urinary bladder). The data was integrated with the SAP-based i.s.h.med system, and the information extraction techniques were applied for textual discharge letters from the respective patients; SONO-CONSULT was used for documentation. By integrating different data sources into the warehouse it is possible to measure the conformity of sonographic results with other methods or inputs. In our evaluations, we applied computer-tomography diagnoses and additional billing diagnoses (from the hospital information system) as a gold-standard. Table <ref type="table" target="#tab_1">1</ref> shows the correlation of SONOCONSULT based diagnosis with CT/MR, diagnoses listed in the discharge letter and diagnoses contained in the hospital information system for a selection of cases from a certain examiner. It was quite interesting that the conformity between SONOCONSULT based diagnoses with the diagnoses contained in the hospital information system was relatively low. Evaluating this issue it was obvious that various diagnosis were not listed in the hospital information system because they were not revenue enhancing and not relevant for all clinical situations. Therefore, we looked at the accordance with the discharge letters which were found to be highly concordant at least for the diagnosis of liver metastasis. Liver cirrhosis is more awkward to detect using ultrasound and has to be in a more advanced stage. Therefore, some of the discharge diagnoses "liver cirrhosis" were only detected using histology or other methods.</p><p>In some cases, there are discrepancies with respect to the formalized background knowledge that still persist after refinement of the rules and checking the data sources. In such cases, explanation-aware mining and analysis components provide appropriate solutions for resolving conflicts and inconsistencies. By supporting the user with appropriate justifications and explanations, misleading patterns can be identified, and the background knowledge can be adapted. The decision whether the background knowledge needs to be adapted is performed by the domain specialist. As we have described in <ref type="bibr" target="#b13">[12]</ref> there are several continuous explanation dimensions in the context of data mining and analysis, that can be utilized for improving the explanation capabilities. In the medical domain, for example, patterns are usually first assessed on the abstract level, before they are checked and verified on concrete patient records, i.e., on a very detailed level of abstraction. Then, discrepancies are modeled in the background knowledge, for example, certain exception conditions for certain subgroups of patients.</p><p>The validation phase is performed on several levels: On the first level, we can use a (partial) gold-standard both for checking the data integration and information extraction tasks. We only require a partial gold-standard, i.e., a sample of the correct relations, because we need to test the functional requirements of the data integration and extraction phases. On the next level, we can incrementally validate the integrated data using the extracted information, or vice versa, using the mined patterns. In the case of discrepancies, we can rely on the partial gold-standard data for verification, or we can identify potential causes and verify these on concrete cases. Therefore, the final decision for the refinements relies on the user, which reviews all proposed refinements in a semi-automatic approach.</p><p>For the refinement steps, we can either extend the (partial) gold-standard, or we perform a boot-strapping approach, using a small gold-standard sample of target concepts for validation, e.g., for validating and refining the information extraction approach, which is in turn used for the validation of the data sources. In the next step, the validation targets can be extended and the process for refinement is applied inversely. The boot-strapping approach for validation and refinement is thus similar to the idea of co-training, e.g., <ref type="bibr" target="#b14">[13]</ref> in machine learning that also starts with a small labeled (correct) dataset and iteratively adapts the models using another co-trained dataset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">CONCLUSIONS</head><p>This paper presented an approach for the validation of mixed-structured data using information extraction and pattern mining methods. In an incremental approach, data can both be validated and refined with an increasing level of accuracy. The presented approach has been successfully implemented in a medical project targeted at integrating data from clinical information systems, documentation systems, and textual discharge letters.</p><p>The experiences and results so far demonstrate the flexibility and effectiveness of the pattern mining and information extraction methods for the presented validation and refinement approach. Having different data sources in the warehouse it is possible to measure the conformity of sonographic results with other methods or inputs. Figure <ref type="figure" target="#fig_0">1</ref> shows the correlation of SONOCONSULT based diagnosis with CT/MR, diagnoses listed in the discharge letter and diagnoses nursed in the hospital information system for a first number of cases. It was quite interesting that the conformity between SONOCONSULT based diagnoses with the diagnoses listed in the hospital information system was quite low. Evaluating this issue it was obvious that various diagnosis were not listed in the hospital information system because they were not revenue enhancing. Therefore, we looked at the accordance with the discharge letters which were found to be highly concordant at least for the diagnosis of liver metastasis.</p><p>Liver cirrhosis is more awkward to be diagnosed with ultrasound and has to be in a more advanced stage. Therefore, some of the discharge diagnoses "liver cirrhosis" were found with histology or other methods. In one case liver cirrhosis was listed in the hospital information system but was neither found with ultrasound nor in the discharge letter. It came out that the input was performed by another department (neurology).</p><p>Within the limited number of examined cases we found only one case of pancreatic mass which was found in the ultrasound examination and listed in the discharge letter. However, it was not included in the hospital information system.</p><p>The first results of the correlations of diagnoses which were input by various sources show that there is a promising high conformity between SonoConsult and discharge letters, but for further quality improvement the correlation with other imaging techniques is very important. With a higher number of cases it </p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Process Model: Validation of Mixed-Structured Data using Pattern Mining and Information Extraction</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1 .</head><label>1</label><figDesc>Exemplary study for a selection of cases concerning liver examinations performed by a certain examiner: Conformity of system diagnoses with various sources of diagnosis input. The columns indicate the degree of correlation of the different sources with SONOCONSULT diagnoses measured by the number of covered cases.</figDesc><table /></figure>
		</body>
		<back>

			<div type="availability">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Data Sources Documents Data Integration Data Integration Pattern Mining System Data Integration Information Extraction</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title/>
		<author>
			<persName><surname>References</surname></persName>
		</author>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Application and Evaluation of a Medical Knowledge-System in Sonography (Sono-Consult)</title>
		<author>
			<persName><forename type="first">Frank</forename><surname>Puppe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Martin</forename><surname>Atzmueller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Georg</forename><surname>Buscher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Matthias</forename><surname>Huettig</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hardi</forename><surname>Lührs</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hans-Peter</forename><surname>Buscher</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. 18th Europ. Conf. on Artificial Intelligence (ECAI 2008)</title>
				<meeting>18th Europ. Conf. on Artificial Intelligence (ECAI 2008)</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="683" to="687" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Medical Data Mining: Knowledge Discovery in a Clinical Data Warehouse</title>
		<author>
			<persName><forename type="first">Jonathan</forename><forename type="middle">C</forename><surname>Prather</surname></persName>
		</author>
		<author>
			<persName><forename type="first">David</forename><forename type="middle">F</forename><surname>Lobach</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Linda</forename><forename type="middle">K</forename><surname>Goodwin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Joseph</forename><forename type="middle">W</forename><surname>Hales</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Marvin</forename><forename type="middle">L</forename><surname>Hage</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">Edward</forename><surname>Hammond</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. AMIA Annual Fall Symposium (AIMA-1997)</title>
				<meeting>AMIA Annual Fall Symposium (AIMA-1997)</meeting>
		<imprint>
			<date type="published" when="1997">1997</date>
			<biblScope unit="page" from="101" to="105" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">CRISP-DM: Towards a Standard Process Model for Data Mining</title>
		<author>
			<persName><forename type="first">Rüdiger</forename><surname>Wirth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jochen</forename><surname>Hipp</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. 4th Intl. Conf. on the Practical Application of Knowledge Discovery and Data Mining. 2000</title>
				<meeting>4th Intl. Conf. on the Practical Application of Knowledge Discovery and Data Mining. 2000</meeting>
		<imprint>
			<publisher>Morgan Kaufmann</publisher>
			<biblScope unit="page" from="29" to="39" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">A Data Warehouse-Based Approach for Quality Management, Evaluation and Analysis of Intelligent Systems using Subgroup Mining</title>
		<author>
			<persName><forename type="first">Martin</forename><surname>Atzmueller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Stephanie</forename><surname>Beer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Frank</forename><surname>Puppe</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. 22nd International Florida Artificial Intelligence Research Society Conference (FLAIRS), accepted</title>
				<meeting>22nd International Florida Artificial Intelligence Research Society Conference (FLAIRS), accepted</meeting>
		<imprint>
			<publisher>AAAI Press</publisher>
			<date type="published" when="2009">2009</date>
			<biblScope unit="page" from="372" to="377" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Exploiting Background Knowledge for Knowledge-Intensive Subgroup Discovery</title>
		<author>
			<persName><forename type="first">Martin</forename><surname>Atzmueller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Frank</forename><surname>Puppe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hans-Peter</forename><surname>Buscher</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. 19th Intl. Joint Conference on Artificial Intelligence (IJCAI-05)</title>
				<meeting>19th Intl. Joint Conference on Artificial Intelligence (IJCAI-05)<address><addrLine>Edinburgh, Scotland</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2005">2005</date>
			<biblScope unit="page" from="647" to="652" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">An Algorithm for Multi-Relational Discovery of Subgroups</title>
		<author>
			<persName><forename type="first">Stefan</forename><surname>Wrobel</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc</title>
				<meeting>null</meeting>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<title level="m">European Symposium on Principles of Data Mining and Knowledge Discovery (PKDD-97)</title>
				<meeting><address><addrLine>Berlin</addrLine></address></meeting>
		<imprint>
			<publisher>Springer Verlag</publisher>
			<date type="published" when="1997">1997</date>
			<biblScope unit="page" from="78" to="87" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Explora: A Multipattern and Multistrategy Discovery Assistant</title>
		<author>
			<persName><forename type="first">Willi</forename><surname>Klösgen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Knowledge Discovery and Data</title>
				<editor>
			<persName><forename type="first">Usama</forename><forename type="middle">M</forename><surname>Mining</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Gregory</forename><surname>Fayyad</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Padraic</forename><surname>Piatetsky-Shapiro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Ramasamy</forename><surname>Smyth</surname></persName>
		</editor>
		<editor>
			<persName><surname>Uthurusamy</surname></persName>
		</editor>
		<imprint>
			<publisher>AAAI Press</publisher>
			<date type="published" when="1996">1996</date>
			<biblScope unit="page" from="249" to="271" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">UIMA: An Architectural Approach to Unstructured Information Processing in the Corporate Research Environment</title>
		<author>
			<persName><forename type="first">David</forename><surname>Ferrucci</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Adam</forename><surname>Lally</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Nat. Lang. Eng</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="issue">3-4</biblScope>
			<biblScope unit="page" from="327" to="348" />
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">ClearTK: A UIMA Toolkit for Statistical Natural Language Processing</title>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">V</forename><surname>Ogren</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">G</forename><surname>Wetzler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Bethard</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">UIMA for NLP workshop at Language Resources and Evaluation Conference (LREC)</title>
				<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Rule-Based Information Extraction for Structured Data Acquisition using TextMarker</title>
		<author>
			<persName><forename type="first">Martin</forename><surname>Atzmueller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Peter</forename><surname>Kluegl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Frank</forename><surname>Puppe</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the LWA-2008, Special Track on Knowledge Discovery and Machine Learning</title>
				<meeting>of the LWA-2008, Special Track on Knowledge Discovery and Machine Learning</meeting>
		<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="1" to="7" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Textmarker: A tool for rule-based information extraction</title>
		<author>
			<persName><forename type="first">Peter</forename><surname>Kluegl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Martin</forename><surname>Atzmueller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Frank</forename><surname>Puppe</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. Biennial GSCL Conference 2009, 2nd UIMA@GSCL Workshop</title>
				<meeting>Biennial GSCL Conference 2009, 2nd UIMA@GSCL Workshop</meeting>
		<imprint>
			<publisher>Gunter Narr Verlag</publisher>
			<date type="published" when="2009">2009</date>
			<biblScope unit="page" from="233" to="240" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Ready for the MACE? The Mining and Analysis Continuum of Explaining Uncovered</title>
		<author>
			<persName><forename type="first">Martin</forename><surname>Atzmueller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Thomas</forename><surname>Roth-Berghofer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">30th SGAI International Conference on Artificial Intelligence. Accepted</title>
				<imprint/>
	</monogr>
	<note>AI-2010</note>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Combining Labeled and Unlabeled Data with Co-Training</title>
		<author>
			<persName><forename type="first">Avrim</forename><surname>Blum</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Tom</forename><surname>Mitchel</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">COLT: Proceedings of the Workshop on Computational Learning Theory</title>
				<imprint>
			<publisher>Morgan Kaufmann</publisher>
			<date type="published" when="1998">1998</date>
			<biblScope unit="page" from="92" to="100" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
