<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">UniNE at CLEF 2017: TF-IDF and Deep-Learning for Author Profiling Notebook for PAN at CLEF 2017</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Nils</forename><surname>Schaetti</surname></persName>
							<email>nils.schaetti@unine.ch</email>
							<affiliation key="aff0">
								<orgName type="institution">University of Neuchâtel</orgName>
								<address>
									<addrLine>rue ; Emile Argand</addrLine>
									<postCode>11 2000</postCode>
									<settlement>Neuchâtel</settlement>
									<country key="CH">Switzerland</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">UniNE at CLEF 2017: TF-IDF and Deep-Learning for Author Profiling Notebook for PAN at CLEF 2017</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">ACCE3BC3250059B7B566887750E8E6BD</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T20:27+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This paper describes and evaluates a strategy for author profiling using TF-IDF and a Deep-Learning model based on Convolutional Neural Networks. We applied this strategy to the author profiling task of the PAN17 challenge and show that it can be applied to different languages (English, Spanish, Portuguese and Arabic). As features, we suggest using a simple cleaning method for both models, and for the Deep-Learning model, a matrix of 2-grams of letters with punctuation marks, beginning and ending 2-grams, as features. Applying this strategy, we determine that the TFIDF-based model is the best one for language variety classification and that the Deep-Learning model achieve the highest accuracy on gender classification. The evaluations are based on four tweet collections (PAN AUTHOR PROFILING task at CLEF 2017).</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Today, a large amount of data is produced by web applications based on social contexts, like social networks and blogs, where a variety of contents (e.g. pictures, videos, articles, links, texts) are shared directly from web sites and smartphones. Social networks like Facebook and Twitter allow a new kind of communication based on fast interactions which generate multimedia contents with their own characteristics, that are difficult to compare with traditional texts like essays and articles.</p><p>This rises new questions : can we detect differences in writing styles between men and women, language varieties, age groups or psychological profiles? Theses questions are appealing as answers to new problems created by the age of social networks and blogs, such as fake news, plagiarism and identity theft. The problem of author profiling is therefore of particular interest.</p><p>Moreover, author profiling is becoming more and more important for applications in marketing, security and forensics. For example, in forensic linguistics, one would like to know certain characteristics (gender, age group, socio-cultural background) of an author of harassing messages from its linguistic profile. In marketing, companies and resellers would like to know the characteristics of people liking or disliking their products based on the analysis of blogs and product reviews.</p><p>This paper is organised as follow. Section 2 introduces the dataset used for training and testing, as well as the methodology used to evaluate our approach. Section 3 describes the cleaning and tokenization process. Section 4 explains the proposed TFIDFbased model. Section 5 describes the Deep-Learning based classifier. In section 6, we evaluate the strategy we created and compare results on the four different test collections. In the last section, we draw conclusions on the main findings and possible future improvements.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Tweet collections and methodology</head><p>To carry out experiments on the author profiling task with different algorithms, we need a common ground composed of the same datasets and evaluation measures. In order to create this common ground, and to allow the large study in the domain of author profiling, the PAN CLEF evaluation campaign was launched ( <ref type="bibr" target="#b6">[7]</ref>). Multiple research groups with different backgrounds from around the world have proposed a profiling algorithm to be evaluated in the PAN CLEF 2017 campaign with the same methodology <ref type="bibr" target="#b4">[5]</ref>.</p><p>All teams have used the TIRA platform to evaluate their strategy. This platform can be used to automatically deploy and evaluate a software <ref type="bibr" target="#b3">[4]</ref>. The algorithms are evaluated on a common test dataset and with the same measures, but also on the base of the time need to produce the response. The access to this test dataset is restricted so that there is no data leakage to the participants during a software run.</p><p>For the PAN CLEF 2017 evaluation campaign, four test collections of tweets were created, one for each of the following languages : English, Spanish, Portuguese and Arabic. Based on these collections, the problem to address was to predict the author's language variety (varieties of the main language) and its gender <ref type="bibr" target="#b5">[6]</ref>.</p><p>The training data was collected from Twitter. For each tweet collections, the texts come from the same language and are composed of tweets from authors, 100 tweets per authors. For each author, there is two labels we can predict :</p><p>1. The author's gender (male, female); 2. The author's language variety, specific to the language;</p><p>The test sets are also texts collected from Twitter and the task is therefore to predict the gender and language variety for each Twitter author in the test data. There is one collection per language (English, Spanish, Portuguese and Arabic). The English collection is composed of 3600 authors, coming from six different countries, United-States, Great Britain, Ireland, New Zealand, Australia and Canada, 600 for each variety, and 1800 for each gender, for a total of 360'000 tweets.</p><p>The Spanish collection is composed of 4200 authors, coming from seven different countries, Colombia, Argentina, Spain, Venezuela, Peru, Child and Mexico, 2100 for each gender, for a total of 420'000 tweets.</p><p>The Portuguese collection is composed of 1200 authors coming from Brazil and Portugal, 600 for each gender, for a total of 120'000 tweets.</p><p>Finally, the Arabic collection is composed of 2400 authors from Gulf, Levantine, Maghrebi and Egypt, 1200 for each gender, for a total of 240'000 tweets.</p><p>An overview of these collections is depicted in the training set is given under the label "Authors" and the total number of tweets in the collection is indicated under the label "Tweets". The label "Languages varieties" shows the varieties for each collections, and the label "Genders" indicates the number of authors for each gender.</p><p>The training data set is well balanced as for each collection, there is the same number of authors for each language variety and gender. The Spanish collection is the biggest with 4'200 authors, and the smallest is the Portuguese collection with 1'200 authors. All the collections have the same number of authors for each language varieties (600), but varies for the genders.</p><p>A similar test set will be used to compare the participants' strategies of the PAN CLEF 2017 campaign, and we don't have information about its size due to the TIRA system.</p><p>For the PAN CLEF 2017 campaign, the software must provide its answer to each problem as an XML data. The response for the gender is a binary choice (male / female), and the language variety is one of the possible outputs for the language of the collection.</p><p>The overall performance of the system is the joint accuracy of the gender and language variety. The accuracy is the number of authors where both the gender and language variety is correctly predicted for the same author divided by the number of authors in the collection.</p><p>The accuracy for language varieties and genders are also computed as the number of correct answers divided by the total number of authors.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Cleaning and Tokenization</head><p>Before selecting the features from the texts, we need to clean the text and extract tokens. This section aims to explain these two steps.</p><p>To carry out these two steps, we apply a serie of rules to the tweet's text in the following order :</p><p>1. Remove URL (http://.../ ); 2. Remove Twitter quotes (@username); 3. Remove special characters #, $ and S; 4. Clean numbers (100.00 → 100 S 00, 100,00 → 100 $ 00, 100'000 → 100 # 000); 5. Tokenize punctuation (???? → " ? ? ? ? "); 6. Remove new line character; 7. Replace useless characters by space (-, ..., *, /, +, ∖); 8. Remove multiple spaces;</p><p>The step 4 aims to introduce three tokens (S, $ and #) that indicate the way the user represents numbers (points or comma for float, comma or apostrophe for thousands).</p><p>Each collection are from a different language and, therefore, may use a different alphabet. For each collection, we keep in the text only the letters and punctuation corresponding to the language's alphabet. The alphabets for each language are depicted in the table 2.</p><p>We keep letters with accents that are not often used in the language to represent their usage by the authors. For example, we keep accents in the English texts to represent the usage of French words by English authors that could help as information for profiling the author. We make the hypothesis that authors use more or less French or Spanish words depending on the country from which they're coming.</p><p>Concerning the hashtags, we keep them and compute them as normal words. At the end of the cleaning process, the words are splitted by space and used as tokens for the feature selection.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">TFIDF-based model</head><p>The TF-IDF (Term-Frequency-Inverse Document Frequency) is a weighting method often used in information retrieval and mostly in text mining. This statistical measure makes it possible to assess the importance of a term in a document, in relation to a collection or corpus. The raw frequency of a term in simply the number of occurrences of this term in a specific document, this frequency is also called term frequency.</p><p>The inverse document frequency is a measure of the importance of a term in the whole collection. In the TF-IDF model, it gives more weight to less frequent terms, considered to be more discriminatory. It consists in calculating the base-10 logarithm of the inverse of the proportion of documents in the corpus that contain the term.</p><p>To describe our problem more formally, a document 𝑑 in our collection is the set of all tweets belonging to a class to predict (gender or language variety). 𝐷 is the set of all documents in the collection and |𝐷| is the number of documents in the collection (|𝐷| = 2 if the problem is to predict gender).</p><p>The term frequency for a term 𝑡 and a document 𝑑 is therefore defined by</p><formula xml:id="formula_0">𝑡𝑓 𝑑,𝑡 = 𝑛 𝑑,𝑡 |𝑑|</formula><p>where 𝑛 𝑑,𝑡 is the number of occurrences of the term 𝑡 in the document 𝑑. The term frequency 𝑡𝑓 𝑑,𝑡 is then the number of occurrence of the term 𝑡 in document 𝑑 divided by the total number of tokens in the document.</p><p>The inverse document frequency of a term 𝑡 in the whole collection is, For each document 𝑑, we compute a vector 𝑡𝑓 𝑖𝑑𝑓 𝑑 with the 𝑡𝑑𝑖𝑑𝑓 values for each Fig. <ref type="figure">1</ref>: Structure of the input features matrix. From left to right, the matrix represents the ratios of 2-grams of letters and of single-letter tokens, the ratio of word ending 2grams, the ratio of word beginning letters, the ratio of punctuation marks, and the ratio of word beginning 2-grams. terms in the collection. If a term does not appear in the document, its value is set to zero. When we want to predict the class (gender or language variety) of a previously unseen author from the collection, we consider it as query 𝑞 and compute the cosinus similarity between the 𝑡𝑓 𝑖𝑑𝑓 𝑑 vector and the vector 𝑡𝑓 𝑞 of term frequencies in 𝑞 :</p><formula xml:id="formula_1">𝑖𝑑𝑓 𝑡 = 𝑙𝑜𝑔 |𝐷| |{𝑑 : 𝑡 ∈ 𝑑}|</formula><formula xml:id="formula_2">𝑠𝑖𝑚(𝑑, 𝑞) = 𝑡𝑓 𝑖𝑑𝑓 𝑑 • 𝑡𝑓 𝑞 ‖𝑡𝑓 𝑖𝑑𝑓 𝑑 ‖ * ‖𝑡𝑓 𝑞 ‖ where 𝑡𝑓 𝑞,𝑡 = 𝑛 𝑞,𝑡 |𝑑|</formula><p>and finally, we choose the predicted class ĉ𝑞 for the query 𝑞 with the biggest similarity. For example, in the case of gender, we choose ĉ𝑞 as ĉ𝑞 = max 𝑑∈{𝑚𝑎𝑙𝑒,𝑓 𝑒𝑚𝑎𝑙𝑒} 𝑠𝑖𝑚(𝑑, 𝑞)</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Two-grams of letters-based Convolutional Neural Networks</head><p>In machine learning, a Convolutional Neural Network (or CNN) is a kind of feedforward artificial neural network <ref type="bibr" target="#b0">[1]</ref> [2] <ref type="bibr" target="#b2">[3]</ref>, in which the patterns of connection between the neurons are inspired from the visual cortex.</p><p>In our system, we applied a CNN to a matrix representing the 2-grams of letters for an author in a collection. The figure <ref type="figure">1</ref> shows the structure of the 2-gram matrix.</p><p>There is a row for each letter in the alphabet. From left to right, the matrix is composed of a first part where ratio of each 2-gram of letters in the author's tweets, the upper line representing the ratio of alone letters. The second part is composed of ratio of ending letters. The next part is the ratio of ending 2-gram of letters computed from the tokens obtained after the cleaning process of section 3.</p><p>The fourth part is the ratio of first letters in tokens, and the fifth the ratio of punctuation marks. Finally, the last part is the ratio of 2-grams found at the beginning of each tokens. This matrix representing an author is the input for the Convolutional Neural Network shown in figure <ref type="figure" target="#fig_1">2</ref>.</p><p>The first layer is a convolution layer of 10 kernels of size 5 × 5. This layer is the input for a second convolution layer of 20 kernels of size 5 × 5 followed by a drop-out layer. This serves as an input for two linear layers based on ReLU. Finally, the outputs are obtained from a softmax function and give the predicted class of the author. The predicted class is then the class with the highest corresponding output of the softmax function.</p><p>The training phase consists of using 90% of the training set for training and 10% to evaluate the performances at each iteration. The figure <ref type="figure" target="#fig_2">3</ref> shows the evolution of training and test losses after each iteration of the training phase for each language collection. Vertical lines show the lowest test loss for each collection.</p><p>For English, the lower loss is attained after 64 iterations and 66 for Spanish. For Portuguese and Arabic, the lower losses are respectively attained after 87 and 38 iterations. We can see that our CNN model quickly overfit, especially for the Arabic language collection. The main challenge with this model is then to fight effectively overfitting.</p><p>At the end of the training phase, we choose the CNN obtained at the iteration with the lowest loss.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Evaluation</head><p>To evaluate our two models we tested their accuracy on each language training collections. The table <ref type="table" target="#tab_3">3</ref> shows the results of 10-fold cross validation for each combination For English, the TF-IDF model attains 83% and 68% accuracy on language variety and gender respectively against 16% and 50% for the random classifier. The CNN model got respectively 65% and 78% accuracy for language variety and gender. The combination of the two models achieve a final accuracy of 65%.</p><p>For Spanish, the TF-IDF model got 93% and 64% accuracy on language variety and gender respectively against 14% and 50% for the random classifier. The CNN model achieves respectively 78% and 72% accuracy for language variety and gender. The combination of the two models got a final accuracy of 67% against 7% for the random classifier.</p><p>For Portuguese, the TF-IDF model achieved an accuracy of 99% and 73% on the language variety and gender classification task respectively. The CNN model got an accuracy of respectively 98% and 85% for the language variety and gender tasks against 50% for a random classifier.</p><p>For Arabic, the TF-IDF model achieved an accuracy of 86% and 68% for the language variety and gender classification problem compared to 25% and 50% for a random classifier. The CNN model got an accuracy of respectively 67.5% and 75% for language variety and gender. The combination of both models gives a final accuracy of 64%.</p><p>We can see that the no free-lunch principle applies here as the TF-IDF achieved better results on the language variety profiling while the CNN model achieves better results when it comes to profiling gender.</p><p>The TF-IDF model achieves an impressive results for the language variety task on the Spanish collection with 93% accuracy while there is 7 different possible varieties (Argentina, Spain, Chile, Mexico, Colombia, Venezuela, Peru). Its best accuracy is at- speaking authors are the easier to profile with an overall accuracy of 83%, a little bit lower than with the 10-fold CV (84%). The Arabic are the hardest to profile with an overall accuracy of 69% against 64% on the 10-fold CV. The language variety accuracy is much higher than the gender accuracy, as for the 10-fold CV.</p><p>The difference between the training results and the previous 10-fold cross validation on the gender problem is 11.8 for English, 8.4 for Spanish, -0.78 for Portuguese and 4.5 for Arabic. We have clear differences between the test and training accuracy except for Portuguese. The table <ref type="table" target="#tab_5">5</ref> shows the results obtained on the four test collections, thanks to the TIRA platform. For the English language collection, the accuracy goes from 83% with the 10-Fold CV to 81.5% (-1.83) for language variety, and from 78% to 75% (-2.86) for gender classification.</p><p>The accuracy on the Spanish language collection goes from 93.23% to 93.36% (-0.13) and from 72.3% to 71.07% (-1.31) for language variety and gender respectively. For the Portuguese collection, the accuracy goes from 99.25% to 98.38% (-0.87) and from 85% to 72% (-12.5) for language variety and gender respectively. Finally, for the Arabic collection, the accuracy goes respectively from 86% to 81% (-4.78) and from 75% to 71% (-3.56) for language variety and gender.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7">Conclusion</head><p>This paper proposes a combination of TFIDF-based model and Deep-Learning Convolutional Neural Network to predict the language variety and gender of Twitter authors. Based on the hypothesis that an author's writing style can be used to extract its country of origin and its gender, we introduced classifiers that can effectively predict these two characteristics. The TFIDF-based model shows a good performance on language variety classification, on the other hand, the CNN model is effective to classify authors The TFIDF-based model performs well on language variety classification and achieves its best performance, on the test dataset, on the Portuguese collection with 98% accuracy, and an interesting accuracy of 93% on the Spanish collection. The performances on the English and Arabic collections stay behind with respectively 81.5% and 81.3%. The CNN model achieves its best performance on the English collection with 75.1% accuracy. Furthermore, we see two ways to improve this strategy in the future. First, the CNN classifier shows signs of overfitting and a great difference appears between the 10-fold CV and the final test results. Some improvements could probably be done on the training phase. Secondly, the matrix of 2-grams could be improved by determining which features are useful or not, this could significantly lower the computational complexity of the model. The biggest challenge for the CNN model is the small size of the training collections and more work could be done on this point to improve the overall performances.</p><p>The biggest challenge of this year's PAN author profiling task were the gender classification problem were our model achieve an average of 72.5% accuracy compare with 88.6% for language variety classification.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>where</head><label></label><figDesc>|𝐷| is the number of classes in the classification problem and |{𝑑 : 𝑡 ∈ 𝑑}| is the number of document(s) where the term 𝑡 appears. The final tfidf value for a document 𝑑 and term 𝑡 is defined by 𝑡𝑓 𝑖𝑑𝑓 𝑑,𝑡 = 𝑡𝑓 𝑑,𝑡 * 𝑖𝑑𝑓 𝑡 = 𝑛 𝑑,𝑡 |𝑑| * 𝑙𝑜𝑔 |𝐷| |{𝑑 𝑗 : 𝑡 ∈ 𝑑 𝑗 }|</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Fig. 2 :</head><label>2</label><figDesc>Fig. 2: Structure of the Convolutional Neural Network on 2-grams of letters with the following layers : 10x5x5 kernel, 20x5x5 kernel, drop-out, two linear layer (ReLU), softmax.</figDesc><graphic coords="6,141.56,455.06,332.23,162.04" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Fig. 3 :</head><label>3</label><figDesc>Fig. 3: Training and test loss after 𝑛 iterations for the gender task on each collection</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>table 1</head><label>1</label><figDesc></figDesc><table><row><cell>. The number of authors from</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1 :</head><label>1</label><figDesc>PAN CLEF 2017 training corpora statistics</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 2 :</head><label>2</label><figDesc>Language alphabet for each language collection</figDesc><table><row><cell>Language Alphabet</cell></row></table><note>English Aa Àà Áá Ââ ÃãBbCcC ¸c ¸DdEe Éé Èè ÊêFfGgHhIi Íí Ïï ÎîJjKkLlMmNn Oo Óó Ôô ÕõPpQqRrSsTtUu Úú Üü ÛûVvWwXxYyZz ?.!,;: Spanish Aa Àà Áá Ââ ÃãBbCcC ¸c ¸DdEe Éé Èè ÊêFfGgHhIi Íí Ïï ÎîJjKkLlMmNn ÑñOo Óó Ôô ÕõPpQqRrSsTtUu Úú Üü ÛûVvWwXxYyZz ?¿.!¡,;: Portuguese Aa Àà Áá Ââ ÃãBbCcC ¸c ¸DdEe Éé Èè ÊêFfGgHhIi Íí Ïï ÎîJjKkLlMmNn Oo Óó Ôô ÕõPpQqRrSsTtUu Úú Üü ÛûVvWwXxYyZz ?.!,;: Arabic</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 3 :</head><label>3</label><figDesc>10-fold cross validation on the four training collections tained on the Portuguese collection with 99% accuracy and its lowest on the gender classification on the Spanish collection. The CNN model achieves its best performance on the Portuguese collection for language variety classification with 98.3% accuracy and its lowest on the English collection for language variety classification with 65.6%.The table4shows the results on the six training collections obtained on the TIRA platform.The results shows the same pattern as the previous 10-fold cross validation. Portuguese-</figDesc><table><row><cell>Corpus</cell><cell>TF-IDF</cell><cell>CNN</cell><cell>Final</cell><cell>Random</cell></row><row><cell>English varieties</cell><cell>0.8333</cell><cell>0.6563</cell><cell></cell><cell>0.1666</cell></row><row><cell>English genders</cell><cell>0.6805</cell><cell>0.7803</cell><cell></cell><cell>0.5000</cell></row><row><cell></cell><cell>0.4724</cell><cell>0.5228</cell><cell>0.6502</cell><cell>0.0833</cell></row><row><cell>Spanish varieties</cell><cell>0.9323</cell><cell>0.7804</cell><cell></cell><cell>0.1428</cell></row><row><cell>Spanish genders</cell><cell>0.6491</cell><cell>0.7238</cell><cell></cell><cell>0.5000</cell></row><row><cell></cell><cell>0.6051</cell><cell>0.5648</cell><cell>0.6747</cell><cell>0.0714</cell></row><row><cell>Portuguese varieties</cell><cell>0.9925</cell><cell>0.9833</cell><cell></cell><cell>0.5000</cell></row><row><cell>Portuguese genders</cell><cell>0.7317</cell><cell>0.8500</cell><cell></cell><cell>0.5000</cell></row><row><cell></cell><cell>0.5313</cell><cell>0.8358</cell><cell>0.8436</cell><cell>0.2500</cell></row><row><cell>Arabic varieties</cell><cell>0.8609</cell><cell>0.6750</cell><cell></cell><cell>0.2500</cell></row><row><cell>Arabic genders</cell><cell>0.6888</cell><cell>0.7500</cell><cell></cell><cell>0.5000</cell></row><row><cell></cell><cell>0.5929</cell><cell>0.5028</cell><cell>0.6456</cell><cell>0.1250</cell></row><row><cell>Overall</cell><cell>0.6228</cell><cell></cell><cell>0.7035</cell><cell>0.3824</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 4 :</head><label>4</label><figDesc>Evaluation for the four training collections</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 5 :</head><label>5</label><figDesc>Evaluation for the four test collections in gender classes. For both model and for all language, we proposed a simple cleaning process used by both classifiers, and we selected features as a matrix of ratio of various 2-grams for the CNN classifier.</figDesc><table><row><cell>Corpus</cell><cell>Variety</cell><cell>Gender</cell><cell>Both</cell><cell>Random</cell></row><row><cell>English</cell><cell>0.8150</cell><cell>0.7517</cell><cell>0.6133</cell><cell>0.0833</cell></row><row><cell>Spanish</cell><cell>0.9336</cell><cell>0.7107</cell><cell>0.6657</cell><cell>0.0714</cell></row><row><cell>Portuguese</cell><cell>0.9838</cell><cell>0.7250</cell><cell>0.7138</cell><cell>0.2500</cell></row><row><cell>Arabic</cell><cell>0.8131</cell><cell>0.7144</cell><cell>0.5863</cell><cell>0.1250</cell></row><row><cell>Overall</cell><cell>0.8863</cell><cell>0.7254</cell><cell>0.6447</cell><cell>0.3824</cell></row></table></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m" type="main">Multi-column deep neural networks for image classification</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">C</forename><surname>Ciresan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Meier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Schmidhuber</surname></persName>
		</author>
		<idno>CoRR abs/1202.2745</idno>
		<ptr target="http://arxiv.org/abs/1202.2745" />
		<imprint>
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Neocognitron: A self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in position</title>
		<author>
			<persName><forename type="first">K</forename><surname>Fukushima</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Biological cybernetics</title>
		<imprint>
			<biblScope unit="volume">36</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="193" to="202" />
			<date type="published" when="1980">1980</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Gradient-based learning applied to document recognition</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Lecun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Bottou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Bengio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Haffner</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE</title>
				<meeting>the IEEE</meeting>
		<imprint>
			<date type="published" when="1998">1998</date>
			<biblScope unit="volume">86</biblScope>
			<biblScope unit="page" from="2278" to="2324" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Improving the Reproducibility of PAN&apos;s Shared Tasks: Plagiarism Detection, Author Identification, and Author Profiling</title>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Gollub</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Rangel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Stamatatos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Information Access Evaluation meets Multilinguality, Multimodality, and Visualization. 5th International Conference of the CLEF Initiative (CLEF 14</title>
				<editor>
			<persName><forename type="first">E</forename><surname>Kanoulas</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Lupu</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Clough</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Sanderson</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Hall</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Hanbury</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">E</forename><surname>Toms</surname></persName>
		</editor>
		<meeting><address><addrLine>Berlin Heidelberg New York</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2014-09">Sep 2014</date>
			<biblScope unit="page" from="268" to="299" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Overview of PAN&apos;17: Author Identification, Author Profiling, and Author Obfuscation</title>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Rangel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Tschuggnall</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Stamatatos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Experimental IR Meets Multilinguality, Multimodality, and Interaction. 8th International Conference of the CLEF Association, CLEF 2017</title>
				<editor>
			<persName><forename type="first">G</forename><surname>Jones</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Lawless</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Gonzalo</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Kelly</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Goeuriot</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><surname>Mandl</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Cappellato</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<meeting><address><addrLine>Dublin, Ireland; Berlin Heidelberg New York</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2017-09-14">Septembre 11-14, 2017. Sep 2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<author>
			<persName><forename type="first">F</forename><surname>Rangel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<title level="m">CLEF 2017 Labs Working Notes</title>
				<editor>
			<persName><forename type="first">L</forename><surname>Cappellato</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Goeuriot</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><surname>Mandl</surname></persName>
		</editor>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Overview of the 4th author profiling task at pan 2016: cross-genre evaluations</title>
		<author>
			<persName><forename type="first">F</forename><surname>Rangel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Verhoeven</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Daelemans</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes Papers of the CLEF</title>
				<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
