<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Context-based Information Classification on Hungarian Invoices</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Gábor</forename><surname>Szegedi</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Bajdikné</forename><surname>Diána</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Imre</forename><surname>Veres</surname></persName>
							<email>veresdia91@gmail.com</email>
						</author>
						<author>
							<persName><forename type="first">Tomáš</forename><surname>Lendák</surname></persName>
							<email>lendak@inf.elte.hu</email>
						</author>
						<author>
							<persName><surname>Horváth</surname></persName>
						</author>
						<author>
							<affiliation key="aff0">
								<orgName type="department" key="dep1">Department of Data Science and Engineering ELTE</orgName>
								<orgName type="department" key="dep2">Faculty of Informatics Budapest</orgName>
								<orgName type="institution">Eötvös Loránd University</orgName>
								<address>
									<postCode>H-1117</postCode>
									<settlement>Budapest</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff1">
								<orgName type="institution">Pázmány Péter sétány 1/C</orgName>
								<address>
									<country key="HU">Hungary</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Context-based Information Classification on Hungarian Invoices</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">3AAFABF27DE08B2E916B1D70ABF8A345</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-25T08:52+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The field of Artificial Intelligence has always strove towards solving problems with computers that were previously only solvable by humans. An interesting challenge we have these years is extracting information from printed documents. In this publication we focus on the sub-domain of classifying pieces of information on printed invoices. Our goal here was to create a solution capable of finding information on scanned invoices without knowing the template of the invoice. The template-less design is important as invoices can have many different structure based on the issuer. First we feed the invoice image to a commercially available Optical Character Recognition (OCR) engine which returns the extracted texts with their bounding boxes. This information itself wouldn't be enough so we enrich it with feature engineering. The engineered features give information about the content of the text and the context of the surrounding of the bounding box as well as meta information about the entire document. The main novelty of our work comes from how we store contextual information. We then classify the text fragments into 8 categories: Invoice Number, Issue Date, Transaction Date, Due Date, Seller tax number, Customer tax number, Total gross and Other. In this paper we measure the performance of several classification algorithms for our data. We achieved 0.93 macro average F1 measure with our best model.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Information extraction can be a very easy or an impossibly hard problem depending on the kind of data we have. In this paper we are dealing with invoices of various sources. We have received thousands of files from our business partner OTP 1 bank to create a solution for their problem of parsing the incoming invoices. The format of the files fall in 3 main categories:</p><p>1. Photographed image of a printed invoice 2. Scanned image of a printed invoice</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Text based PDF invoice</head><p>The problem with the dataset is that we can't use a finite set of templates. The law does not constrain the format Figure <ref type="figure">1</ref>: A sample printed invoice that was photographed using a mobile phone. of invoices and thus each issuer has its own template for generating invoices. Even if we could manage to create a template for each of the issuers, the number of issuers is growing.</p><p>Because of these given conditions we have developed a solution that is generic and free of any premises about the structure of the input document.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.1">Related Work</head><p>The main approaches to invoice recognition are Template based, Graph Convolutional Neural Network (CNN) based and direction based.</p><p>Information Extraction (IE) from structured documents has been a studied problem for decades now. One of the first works was The generic information extraction system by Jerry R. Hobbs <ref type="bibr" target="#b0">[1]</ref>. In his work he laid out an architecture for creating systems capable of extracting information from text using a rule based layer-by-layer approach.</p><p>Parsing invoice data is a sub-domain of IE. In the case when the input data consists of a finite set of known layouts we can simply extract information based on the known positions of the key pieces. Such a problem does not require sophisticated Machine Learning (ML) solutions, but rather a simple algorithmic approach.</p><p>The problem gets significantly harder if we do not know the document templates at training time. In such a case we need to create a generic solution that is smart enough to extract information from unknown document layouts. This field has been studied as well in recent years.</p><p>A very recent study entitled An Invoice Reading System Using a Graph Convolutional Network was published in 2019, in which the authors turned the output of an Optical Character Recognition (OCR) system into a Graph <ref type="bibr" target="#b1">[2]</ref>. The graph's nodes contained the text outputs from the OCR and edges were pointing to nearby nodes based on the location information given by the OCR output. The nodes were then fed to a Graph based CNN to classify for entities of interest.</p><p>CloudScan -A configuration-free invoice analysis system using recurrent neural networks is another article from 2017 in which the authors have created a fully fledged endto-end pipeline for parsing invoices <ref type="bibr" target="#b2">[3]</ref>. They get the text from OCR and create N-grams of words in the same line up to a length of 4. They enrich their entities with features based on the properties of the piece of text in the N-gram. They add contextual features based on the closest 4 entities above, below, left and right of the current text. The classifier they use is a Long Short-Term Memory (LSTM) Recurrent Neural Network <ref type="bibr" target="#b3">[4]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Scope</head><p>Our goal was to find the most relevant information needed by our industry partner. The scope of our work was to extract the following information from the given invoices:</p><p>• Invoice Number: The unique identifier of the invoice;</p><p>• Issue Date: The date the invoice was issued;</p><p>• Transaction Date: The date of reconciliation;</p><p>• Due Date: The date until the invoice has to be reconciled;</p><p>• Seller tax number: The tax number of the seller party;</p><p>• Customer tax number: The tax number of the customer party (This is only included if the customer is a company, and not an individual);</p><p>• Total gross: The total amount to be paid by the customer party.</p><p>The invoices considered are all printed invoices, thus no handwritten invoices were used as those invoices are getting ousted. The invoices can be image based or "text" based such that, e.g. a portable document format (PDF).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Our approach</head><p>In any ML problem the biggest dividing factor is how the data is represented. In our case we have a very high level representation of the data given images of invoices. We didn't find it feasible to train a complex Neural Network that operates directly on the image data for the following reasons:</p><p>1. The dimensionality of the images is large and we can't shrink them without loosing the small but essential textual information.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Training times without dimensionality reduction</head><p>would be too big even on a special hardware.</p><p>3. There are invoices that come in textual PDF format.</p><p>For these 3 reasons we chose to simplify the high level representation of the invoice data. We began by breaking down the invoice into a list of Fragments. A Fragment is a piece of text with its location (X, Y, page) and dimensionality (width, height) information.</p><p>Creating the list of Fragments is a rather simple task. For text PDF invoices the representation is basically the same as we desire. We just need to use a Python API<ref type="foot" target="#foot_0">2</ref> to get the text with the bounding boxes from each page of the PDF. Converting image based invoices to Fragments is also trivial as we can use an OCR API which returns a list of texts with their bounding boxes. We have integrated our pipeline with Tesseract<ref type="foot" target="#foot_1">3</ref> , Azure OCR <ref type="foot" target="#foot_2">4</ref> and OCR Space<ref type="foot" target="#foot_3">5</ref> . We found Tesseract to be quite inaccurate compared to the other two, maybe because our invoices are in Hungarian and not in English. In the end we used OCR Space.</p><p>Once we have the list of Fragments the task is finding the relevant ones and extracting the information out of the text. This can be formulated as a classification task, where each Fragment is a separate input to the classifier and the model assigns a label denoting what kind of information is in the Fragment. For each label type we select the Fragment of the highest probability as the Fragment containing the given piece of information.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Feature Engineering</head><p>It should be possible to build a model that we can feed the entire fragment list as input and ask the model to return what is the most probable fragment for each label is. We could use Recurrent Neural Networks and consider the list of fragments a sequence of input. However this would make the problem very hard.</p><p>Instead of this we have chosen to do extensive feature engineering to add relevant information to each fragment and do classification on each fragment individually. In total we have added 104 features to the existing 6, such that (text, X, Y, width, height, page), what is the main added value of our work. We can group these engineered features into 4 groups which we go into detail below.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Positional features</head><p>We have added a lot of features that are describing the position of the Fragment relative to the document. We intuitively found this to be very important. For example the invoice number is usually on the top of the first page of a document. To describe this first we needed to add a feature for storing which page the Fragment is part of. Then we need the features to describe the X and Y coordinates of the Fragment on the given page. Lastly we have added features to describe the dimensionality of the Fragment.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Document features</head><p>Another set of features added were to describe the entire document. Knowing the positional attributes of a Fragment is not meaningful without knowing the dimensionality attributes of the document. Thus we have added features for document width and height, their ratios, the number of pages. We also added features describing where are the outermost Fragments for the current page and for the entire document.</p><p>Text features In ML, when dealing with text, the problem arises from the fact that our models are only capable of working with numbers. We need to encode the text into a finite set of features all the time, so that is what we did as well for our Fragments.</p><p>One obvious feature that we can always use to quantify text is the length of the text we have. We added this of course to our features.</p><p>Another set of features we used were counters of special characters in the Fragment. For example the count of whitespaces is an important feature telling us how many words are there in the fragment. Counters of digits and alphabetical characters are also good indicators of what kind of text are we dealing with. The count of slashes and dashes could imply that the current Fragment contains a bank account number or an invoice number. We could add many more special characters depending on the data we have.</p><p>Lastly we have added a few Boolean features derived from the text based on whether the text matches a given regular expression or not. We created regular expression features for matching date, tax number, currency and decimal point.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Contextual features</head><p>The feature categories we discussed so far are standard features we have seen in other related works as well <ref type="bibr" target="#b2">[3]</ref> [5] <ref type="bibr" target="#b1">[2]</ref>. These features in themselves wouldn't be enough to do labeling effectively, we need the context of the Fragments as well. We have also seen the idea of adding context information to the Fragments in other works but there is no consensus on what is the best way to do this.</p><p>In <ref type="bibr" target="#b2">[3]</ref>, the authors give context information by adding the features of the nearest neighboring Fragments in the 4 general directions (Up, Down, Left, Right).In <ref type="bibr" target="#b1">[2]</ref>, a very similar method is used as the graph nodes connect to the 4 general directions as well. In <ref type="bibr" target="#b5">[6]</ref>, the authors took a different approach by storing the polar coordinates of each fragment relative to the current fragment.</p><p>We take a new approach in capturing the context of the Fragment. We have identified 38 keywords in invoices that are indicative of important information in the vicinity. For each Fragment we search for the nearest occurrence of all the keywords. We calculate the distance from the Fragment to the keyword and store the normalized x and y values as features for each of the keywords. After doing this for each keyword we get a complex web over our invoice document. We can see this mechanism in Figure <ref type="figure">2</ref> for a single Fragment of a sample invoice.</p><p>Using such keywords comes naturally to us humans when we interpret an invoice. For example if we take a look at an invoice and we see 2 tax numbers we can decide which belongs to the seller by deciding which one is closer to the keyword Seller or Provider 6 .</p><p>Figure <ref type="figure">2:</ref> A sample invoice where the Fragment in the green bounding box is the tax number of the seller and the closest keywords are highlighted with blue bounding boxes. Note that some of the keywords occur multiple times on the invoice but we chose the closest one for the current Fragment. There are 2 special cases to note when pointing the to the nearest keyword. The first one is when the current fragment contains one of the keywords. In that case we chose to set the value of the x and y features to be zero. The second special case is when a certain keyword is not to be found on the invoice. In this case we assign −1 to the x and y features of the keyword. Our approach is different from the related works because we neither restrict the context to the 4 nearest Fragments in the general directions, nor do we include all the features. The contextual information is relevant because of the keywords these vectors are pointing towards.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Classification</head><p>After preparing the data we had to train a classifier for labeling the Fragments. During the training and evaluation we had to be cautious as the dataset is very unbalanced because on each document there are hundreds of text Fragments but we are only interested in 7 of them. See Figure <ref type="figure" target="#fig_0">3</ref> for the distribution of the 8 class labels.</p><p>As 93% of this data was labeled as Other, a classifier that assigns the Other label blindly to any input would achieve an overall accuracy of 93%. This is of course not good as our focus is on the classes with low cardinality. To overcome this we are using the macro average of the F1 Score of the classes to compare the classifiers.</p><p>We have chosen 3 classifiers for comparison: Decision Tree Classifier, Random Forest Classifier 7 and Extreme 7 For the Decision Tree and Random Forest models we used the implementations from scikit-learn Gradient Boosting Classifier <ref type="bibr" target="#b6">[7]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Experiments</head><p>The data we received consisted of thousands of Invoices from different issuers in various file formats. About half of them were electronic invoices in textual PDFs, while other half was images of either photographs or scans of printed invoices.</p><p>First we had to manually annotate the ground truth about the invoices. For each invoice we stored the relevant information pieces that were in scope.</p><p>After that we had run the OCR engine on all of the image based invoices. We then filtered out those image based invoices where the OCR output did not contain at least the invoice number and the seller tax number. We did this because the OCR engine optimization is out of our scope and this way we eliminated any errors due to image quality or bugs in the OCR engine used. We had 2000 invoices that met this criteria.</p><p>Lastly we ran the algorithm to convert the filtered invoices into Fragments. We manually labeled the Fragments to eliminate any OCR issues, or data format differences there may be. This was the data that we could build our models upon.</p><p>We have split the Fragments into train, test and validation sets as usual in the field. The ratios used were 60% for training, 20% for validation (tuning the hyper-parameters) and 20% for comparing the trained and optimized models. The results are visible in Table <ref type="table">5</ref>. Without surprise the XGBoost Classifier performs the best out of these 3, but even the Decision Tree and Random Forest models are very close to the Gradient Boosting model. XGBoost outperforms the other 2 on all of the labels and in the overall macro average as well. The tree based models fall behind in the Total Gross label.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Conclusion and future work</head><p>In this work we presented a method of identifying key information on Invoices. We have built upon external dependencies via Optical Character Recognition, which is used to break down the image based invoice into text pieces with positional information called Fragments. We did an extensive feature engineering so that the Fragments contain features regarding the properties document, the contained text and the context of the text as well.</p><p>The main added value of our work comes from how we add contextual information to the Fragments. It works by pointing vectors from the Fragment to each of the nearest keywords. This gives the power of our method to generalize so that unlike some of the previous works it can perform well on previously unseen Invoices.</p><p>However our work does not stop here. We plan on creating an end-to-end solution for extracting the information from invoices. We have merely identified the Fragments correctly, but extracting the information from the selected Fragments still holds challenges.</p><p>As our main focus was on feature engineering there can still be room for improvement in the modeling side. We are planning on experimenting with Neural Network models in the future to see if we can improve our accuracy.</p><p>We also plan on embedding the system in a service, where the user sends in the invoice and our system either returns the extracted information, or a new version of the invoice image where the key information points are highlighted like in Figure <ref type="figure">4</ref>.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Class distribution of Fragments on one of our test sets.</figDesc><graphic coords="4,56.69,80.50,231.03,238.44" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0"><head></head><label></label><figDesc></figDesc><graphic coords="5,307.56,80.50,231.00,239.72" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 :</head><label>1</label><figDesc>Comparison of models after hyperparameter optimization. The numbers are F1 scores.</figDesc><table><row><cell></cell><cell>Decision</cell><cell>Random</cell><cell>XGBoost</cell></row><row><cell></cell><cell>Tree</cell><cell>Forest</cell><cell></cell></row><row><cell>Other</cell><cell>0.99</cell><cell>0.99</cell><cell>1.00</cell></row><row><cell>Invoice</cell><cell>0.81</cell><cell>0.86</cell><cell>0.89</cell></row><row><cell>number</cell><cell></cell><cell></cell><cell></cell></row><row><cell>Issue date</cell><cell>0.88</cell><cell>0.89</cell><cell>0.93</cell></row><row><cell>Transaction</cell><cell>0.87</cell><cell>0.89</cell><cell>0.94</cell></row><row><cell>date</cell><cell></cell><cell></cell><cell></cell></row><row><cell>Due date</cell><cell>0.88</cell><cell>0.93</cell><cell>0.96</cell></row><row><cell>Seller tax</cell><cell>0.89</cell><cell>0.91</cell><cell>0.92</cell></row><row><cell>number</cell><cell></cell><cell></cell><cell></cell></row><row><cell>Customer</cell><cell>0.92</cell><cell>0.93</cell><cell>0.94</cell></row><row><cell>tax number</cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="2">Total gross 0.71</cell><cell>0.75</cell><cell>0.87</cell></row><row><cell>Macro Av-</cell><cell>0.8685</cell><cell>0.8952</cell><cell>0.9297</cell></row><row><cell>erage</cell><cell></cell><cell></cell><cell></cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_0">We used pdfminer.six API https://pypi.org/project/ pdfminer.six/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_1">https://github.com/tesseract-ocr/tesseract</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_2">https://docs.microsoft.com/en-us/ azure/cognitive-services/computer-vision/ concept-recognizing-text</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_3">https://ocr.space/</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Acknowledgements The research has been supported by the European Union, co-financed by the European Social Fund (EFOP-3.6.2-16-2017-00013, Thematic Fundamental Research Collaborations Grounding Innovation in Informatics and Infocommunications). Supported by Telekom Innovation Laboratories (T-Labs), the Research and Development unit of Deutsche Telekom and EIT Digital.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">The generic information extraction system</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">R</forename><surname>Hobbs</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Fifth Message Understanding Conference (MUC-5): Proceedings of a Conference Held in</title>
				<meeting><address><addrLine>Baltimore, Maryland, August</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1993">25-27, 1993. 1993</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">An invoice reading system using a graph convolutional network</title>
		<author>
			<persName><forename type="first">D</forename><surname>Lohani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Belaïd</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Belaïd</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Asian Conference on Computer Vision</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="144" to="158" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Cloudscan-a configuration-free invoice analysis system using recurrent neural networks</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">B</forename><surname>Palm</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Winther</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Laws</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">14th IAPR International Conference on Document Analysis and Recognition (ICDAR)</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2017">2017. 2017</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="406" to="413" />
		</imprint>
	</monogr>
	<note>Figure 4: The key information pieces our system can identify from the Fragments</note>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Long short-term memory</title>
		<author>
			<persName><forename type="first">S</forename><surname>Hochreiter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Schmidhuber</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Neural computation</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="issue">8</biblScope>
			<biblScope unit="page" from="1735" to="1780" />
			<date type="published" when="1997">1997</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Recognition of ocr invoice metadata block types</title>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">T</forename><surname>Ha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Nevěřilová</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Horák</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Text, Speech, and Dialogue</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="304" to="312" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Field extraction from administrative documents by incremental structural templates</title>
		<author>
			<persName><forename type="first">M</forename><surname>Rusinol</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Benkhelfallah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Poulain</surname></persName>
		</author>
		<author>
			<persName><surname>Dandecy</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">2013 12th International Conference on Document Analysis and Recognition</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2013">2013</date>
			<biblScope unit="page" from="1100" to="1104" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Xgboost: A scalable tree boosting system</title>
		<author>
			<persName><forename type="first">T</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Guestrin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 22nd acm sigkdd international conference on knowledge discovery and data mining</title>
				<meeting>the 22nd acm sigkdd international conference on knowledge discovery and data mining</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="785" to="794" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
