<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">RatVec: A General Approach for Low-dimensional Distributed Vector Representations via Rational Kernels</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Eduardo</forename><surname>Brito</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Fraunhofer Center for Machine Learning</orgName>
								<address>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="institution">Fraunhofer IAIS</orgName>
								<address>
									<addrLine>Schloss Birlinghoven</addrLine>
									<postCode>53757</postCode>
									<settlement>Sankt Augustin</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Bogdan</forename><surname>Georgiev</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Fraunhofer Center for Machine Learning</orgName>
								<address>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="institution">Fraunhofer IAIS</orgName>
								<address>
									<addrLine>Schloss Birlinghoven</addrLine>
									<postCode>53757</postCode>
									<settlement>Sankt Augustin</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Daniel</forename><surname>Domingo-Fernández</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Fraunhofer Center for Machine Learning</orgName>
								<address>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
							<affiliation key="aff2">
								<orgName type="institution">Fraunhofer SCAI</orgName>
								<address>
									<addrLine>Schloss Birlinghoven</addrLine>
									<postCode>53757</postCode>
									<settlement>Sankt Augustin</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
							<affiliation key="aff3">
								<orgName type="department">B-IT</orgName>
								<orgName type="institution">University of Bonn</orgName>
								<address>
									<addrLine>Endenicher Allee 19a</addrLine>
									<postCode>53115</postCode>
									<settlement>Bonn</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Charles</forename><forename type="middle">Tapley</forename><surname>Hoyt</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Fraunhofer Center for Machine Learning</orgName>
								<address>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
							<affiliation key="aff2">
								<orgName type="institution">Fraunhofer SCAI</orgName>
								<address>
									<addrLine>Schloss Birlinghoven</addrLine>
									<postCode>53757</postCode>
									<settlement>Sankt Augustin</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
							<affiliation key="aff3">
								<orgName type="department">B-IT</orgName>
								<orgName type="institution">University of Bonn</orgName>
								<address>
									<addrLine>Endenicher Allee 19a</addrLine>
									<postCode>53115</postCode>
									<settlement>Bonn</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Christian</forename><surname>Bauckhage</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Fraunhofer Center for Machine Learning</orgName>
								<address>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="institution">Fraunhofer IAIS</orgName>
								<address>
									<addrLine>Schloss Birlinghoven</addrLine>
									<postCode>53757</postCode>
									<settlement>Sankt Augustin</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
							<affiliation key="aff3">
								<orgName type="department">B-IT</orgName>
								<orgName type="institution">University of Bonn</orgName>
								<address>
									<addrLine>Endenicher Allee 19a</addrLine>
									<postCode>53115</postCode>
									<settlement>Bonn</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">RatVec: A General Approach for Low-dimensional Distributed Vector Representations via Rational Kernels</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">394D6801D32C0B31D4519D82594340D6</idno>
					<idno type="DOI">10.7910/DVN/JMFHTN</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T18:29+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Representation Learning</term>
					<term>Kernel Principal Component Analysis</term>
					<term>Bioinformatics</term>
					<term>Natural Language Processing</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>We present a general framework, RatVec, for learning vector representations of non-numeric entities based on domain-specific similarity functions interpreted as rational kernels. We show competitive performance using k -nearest neighbors in the protein family classification task and in Dutch spelling correction. To promote re-usability and extensibility, we have made our code and pre-trained models available at https://github.com/ratvec.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>The success of distributed vector representations in natural language processing (NLP) tasks has motivated their use for other areas such as biological sequence analysis <ref type="bibr" target="#b0">[1]</ref>. Most of them rely on the distributional hypothesis <ref type="bibr" target="#b7">[8]</ref>: "words that appear in similar context are similar". We aim to relax this constraint with a general framework to learn vector representations of non-numeric entities including text, DNA sequences, and protein sequences based on similarity functions that can be expressed as rational kernels <ref type="bibr" target="#b4">[5]</ref>. Supported by the robust theoretical framework behind rational kernels, we obtain vector representations via kernel PCA (KPCA) <ref type="bibr" target="#b10">[11]</ref> that, together with a simple k nearest neighbors (k NN) classifier, constitute an efficient and explainable classification pipeline showing competitive performance in two different tasks: protein family classification and Dutch spelling correction. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Related Work</head><p>Our approach extends our previous work on KPCA embeddings, where we studied vector representations for words and DNA sequences via KPCA replacing the dot product by a specific similarity function <ref type="bibr" target="#b2">[3]</ref>. There are also a myriad of previous kernel methods for NLP tasks such as text classification <ref type="bibr" target="#b5">[6,</ref><ref type="bibr" target="#b9">10]</ref> and named entity recognition <ref type="bibr" target="#b6">[7]</ref>. However, they are either restricted to a specific kernel function (i.e. to a concept of similarity) or lack in explicit vector representations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Approach</head><p>Let X be a dataset consisting of n elements, to which we will further refer as the full vocabulary; and k a rational kernel (a particular similarity function). Computing a kernel matrix K from X may be computationally prohibitive for large datasets due to its time and space complexity (O(n 2 )). Hence, we select only the m elements considered "most representative" (in a domain-specific formulation) from X to construct our representative vocabulary V and compute a kernel matrix K V from all element pairs from V . After centering and diagonalizing K V , we construct our projection matrix P V . This is required to generate m-dimensional representations for the full vocabulary. In a last step, we assign the first d ≤ m components of each computed vector to each full vocabulary element. There are the resulting d-dimensional vector representations obtained with our approach. When k is suitable for the task, d being of a lower order of magnitude than m can lead to optimal results, as we will see in Section 4.1. The choice of m can be adjusted to fit the computational resources of the user.</p><p>The produced representations tend to cluster naturally according to the domain-specific concept of similarity applied by the selected rational kernel, as we can see in Fig. <ref type="figure" target="#fig_0">1</ref>. This has two main advantages:</p><p>1. A simple k NN classifier (eventually 1NN) can suffice for classification tasks. 2. Learning the vector representations and k NN constitute an explainable classification pipeline: an entity is assigned to a particular class because it is similar to the labeled entities used to train the classifier. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Protein Family Classification</head><p>In this task, we classify protein sequences to their corresponding families with the same Swiss-Prot dataset as in <ref type="bibr" target="#b0">[1]</ref>, containing 7,027 protein families and 324,018 protein sequences<ref type="foot" target="#foot_0">5</ref> . The system is evaluated for each protein family by 10-fold cross-validation on a balanced dataset consisting of all amino acid sequences of the family and as many randomly sampled sequences from all other families. We produced protein representations with 25 dimensions applying our approach with bigram and trigram similarity and trained a nearest neighbor classifier (1NN). We generated our representative vocabulary by taking the shortest sequence of the 1,000 most frequent protein families. We evaluated this pipeline in the same setting as <ref type="bibr" target="#b0">[1]</ref>. We report the weighted average accuracy results in Table 1. Also, the results of the evaluation limited to subsets of the full dataset are presented (e.g. sequences belonging to the 1,000 most frequent protein families).</p><p>The results from Table <ref type="table" target="#tab_0">1</ref> show that our approach with the bigram similarity reaches the same accuracy as the baseline and or even outperforms it when we restrict the dataset to the first 3,000 or 4,000 most frequent protein families. Considering that the reported BioVec representations are four times longer than ours, our approach to encode proteins seems to be more efficient.</p><p>In contrast to <ref type="bibr" target="#b0">[1]</ref>, we do not train any support-vector machine (SVM) but a 1NN classifier. To assess the impact of the different classification algorithm, we evaluated a pretrained BioVec model<ref type="foot" target="#foot_1">6</ref> with our setup. The results (Table <ref type="table" target="#tab_1">2</ref>), showing that 1NN is more suitable than SVMs for this task, are in line with our previous experiments on classification tasks using distributed vector representations, where simple algorithms such as k NN and logistic regression outperform more complex ones such as random forests or neural networks <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b3">4]</ref>.</p><p>Table <ref type="table">3</ref>. Results of our system compared to Valkuil (CLIN28 shared task baseline) in terms of successful corrections (TP), wrong corrections (FN), and accuracy <ref type="bibr" target="#b1">[2]</ref>  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Dutch Spelling Correction</head><p>We also evaluated our approach by participating in the CLIN28 shared task, where our system obtained the best F1 score among the competing teams <ref type="bibr" target="#b1">[2]</ref>. The task focused on correcting errors in extracts from Dutch Wikipedia pages. We used an older implementation of our approach that we keep for reproducibility 7 .</p><p>Spell checkers involve two steps: misspelling detection and correction. The latter generally requires ranking a set of correction candidates. Generating them implies finding all valid words differing from the detected misspelling less than a determined edit distance, which is mostly set to 1 for real-world applications to limit computation time <ref type="bibr" target="#b11">[12]</ref>. We avoid this strong constraint with our approach.</p><p>In our RatVec framework, our full vocabulary is wdutch, a word list from the OpenTaal project, from which the 3000 most frequent words form our representative vocabulary. The applied rational kernel is the composition of the bigram similarity <ref type="bibr" target="#b8">[9]</ref> with the homogeneous polynomial kernel of degree 2. We generated a vector representation for each full vocabulary word with 2000 dimensions.</p><p>Once a misspelling is detected, we compute its RatVec representation and search for the closest precomputed vector. Its related word (its nearest neighbor) is our correction to the misspelling. Formally, this is equivalent to training a 1NN classifier where each valid word is assigned a different label.</p><p>Our RatVec framework is only relevant during the correction phase for spelling mistakes that are related to the word form. Hence, we restrict our analysis to the correction results on the five error categories where the word form is relevant, namely those displayed in Table <ref type="table">3</ref>. Although the baseline system Valkuil achieves a better average accuracy than our approach for the analyzed misspellings, RatVec outperforms Valkuil in some categories where it completely fails (redundant punctuation errors) or where it cannot be even evaluated because it failed to detect any error (capitalization and archaic spelling errors). From these results, we interpret that our word vectors encode word forms in a suitable way so that similar words can be retrieved.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Conclusion and Future Work</head><p>We showed that our approach involving rational kernels on KPCA for vector space embeddings provides rich representations for two different kinds of en-7 https://github.com/fraunhofer-iais/kpca embeddings tities: proteins and words. In the first application, we presented how to learn protein vectors from amino acid sequences and how to apply them to predict their protein family. In the second, we learned word representations that encode word form information so that they can be applied to correct misspellings once these are detected. In both tasks, our results are comparable to state-of-the-art approaches. Thanks to the simplicity of k NN classifiers and the interpretable similarity concept we apply to generate the vector representations, our approach may be advantageous for some real-world applications compared to more complex machine learning models such as deep neural networks.</p><p>In future work, we will explore possibilities of learning optimal similarity metrics (modeled as transducers) that, incorporated in our presented approach, solve a particular task.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Visualization of learned vectors for two different protein families using bigram similarity and a representative vocabulary of 1,000 sequences. A clear separation can be observed by only using the first three components of the vectors. Best seen in color.</figDesc><graphic coords="2,212.58,115.83,190.21,95.61" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head></head><label></label><figDesc>a . Capitaliz. Non-word Red. punct Mis. punct Archaic Total (165 errors) (62 errors) (53 errors) (18 errors) (5 errors) (303 errors) System TP FN Acc. TP FN Acc. TP FN Acc. TP FN Acc. TP FN Acc. TP FN Acc. Valkuil 0 0.90 24 15 0.62 8 5 0.62 3 3 0.50 1 0 1.0 141 35 0.80 a https://github.com/LanguageMachines/CLIN28 ST spelling correction</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 .</head><label>1</label><figDesc>Accuracy of RatVec in different subdatasets compared to BioVec<ref type="bibr" target="#b0">[1]</ref> </figDesc><table><row><cell>Dataset</cell><cell cols="3">BioVec (baseline) RatVec (bigram sim.) RatVec (trigram sim.)</cell></row><row><cell cols="2">Top 1,000 families 0.94</cell><cell>0.94</cell><cell>0.91</cell></row><row><cell cols="2">Top 2,000 families 0.93</cell><cell>0.93</cell><cell>0.91</cell></row><row><cell cols="2">Top 3,000 families 0.92</cell><cell>0.93</cell><cell>0.91</cell></row><row><cell cols="2">Top 4,000 families 0.91</cell><cell>0.93</cell><cell>0.91</cell></row><row><cell>All families</cell><cell>0.93</cell><cell>0.93</cell><cell>0.91</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 .</head><label>2</label><figDesc>Accuracy of BioVec representations with a SVM<ref type="bibr" target="#b0">[1]</ref> and a k NN classifier</figDesc><table><row><cell cols="5">Top 1000 families Top 2000 families Top 3000 families Top 4000 families All families</cell></row><row><cell>SVM 1NN</cell><cell>SVM 1NN</cell><cell>SVM 1NN</cell><cell>SVM 1NN</cell><cell>SVM 1NN</cell></row><row><cell>0.94 0.94</cell><cell>0.93 0.94</cell><cell>0.92 0.93</cell><cell>0.91 0.93</cell><cell>0.93 0.93</cell></row><row><cell cols="2">4 Experimental Results</cell><cell></cell><cell></cell><cell></cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_0">http://dx.doi.org/10.7910/DVN/JMFHTN</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_1">https://github.com/kyu999/biovec</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Continuous distributed representation of biological sequences for deep proteomics and genomics</title>
		<author>
			<persName><forename type="first">E</forename><surname>Asgari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">R</forename><surname>Mofrad</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">PloS one</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="issue">11</biblScope>
			<biblScope unit="page">e0141287</biblScope>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Detecting and correcting spelling errors in high-quality dutch wikipedia text</title>
		<author>
			<persName><forename type="first">M</forename><surname>Beeksma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Van Gompel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Kunneman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Onrust</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Regnerus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Vinke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Brito</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Bauckhage</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Sifa</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computational Linguistics in the Netherlands Journal</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="page" from="122" to="137" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">KPCA embeddings: an unsupervised approach to learn vector representations of finite domain sequences</title>
		<author>
			<persName><forename type="first">E</forename><surname>Brito</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Sifa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Bauckhage</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">LWDA</title>
		<imprint>
			<biblScope unit="page" from="87" to="96" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Towards German word embeddings: A use case with predictive sentiment analysis</title>
		<author>
			<persName><forename type="first">E</forename><surname>Brito</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Sifa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Cvejoski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Ojeda</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Bauckhage</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Data Science-Analytics and Applications</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="59" to="62" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Rational kernels: Theory and algorithms</title>
		<author>
			<persName><forename type="first">C</forename><surname>Cortes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Haffner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mohri</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Machine Learning Research</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="page" from="1035" to="1062" />
			<date type="published" when="2004-12">Dec 2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Latent semantic kernels</title>
		<author>
			<persName><forename type="first">N</forename><surname>Cristianini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Shawe-Taylor</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Lodhi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Intelligent Information Systems</title>
		<imprint>
			<biblScope unit="volume">18</biblScope>
			<biblScope unit="issue">2-3</biblScope>
			<biblScope unit="page" from="127" to="152" />
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Fine-grained classification of named entities exploiting latent semantic kernels</title>
		<author>
			<persName><forename type="first">C</forename><surname>Giuliano</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Thirteenth Conference on Computational Natural Language Learning</title>
				<meeting>the Thirteenth Conference on Computational Natural Language Learning</meeting>
		<imprint>
			<publisher>Association for Computational Linguistics</publisher>
			<date type="published" when="2009">2009</date>
			<biblScope unit="page" from="201" to="209" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Distributional structure</title>
		<author>
			<persName><forename type="first">Z</forename><forename type="middle">S</forename><surname>Harris</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Word</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="issue">2-3</biblScope>
			<biblScope unit="page" from="146" to="162" />
			<date type="published" when="1954">1954</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">N-gram similarity and distance</title>
		<author>
			<persName><forename type="first">G</forename><surname>Kondrak</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Int. Symp. on String Processing and Information Retrieval</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2005">2005</date>
			<biblScope unit="page" from="115" to="126" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Text classification using string kernels</title>
		<author>
			<persName><forename type="first">H</forename><surname>Lodhi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Saunders</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Shawe-Taylor</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Cristianini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Watkins</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Machine Learning Research</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="419" to="444" />
			<date type="published" when="2002-02">Feb. 2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Kernel principal component analysis</title>
		<author>
			<persName><forename type="first">B</forename><surname>Schölkopf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Smola</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">R</forename><surname>Müller</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International conference on artificial neural networks</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="1997">1997</date>
			<biblScope unit="page" from="583" to="588" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">Context-based spelling correction for the Dutch language: Applied on spelling errors extracted from the Dutch wikipedia revision history</title>
		<author>
			<persName><forename type="first">L</forename><surname>Tijhuis</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
