<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Convolutional Neural Networks for Author Profiling</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Sebastian Sierra</string-name>
          <email>ssierral@unal.edu.co</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Manuel Montes-y-Gómez</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Thamar Solorio</string-name>
          <email>solorio@cs.uh.edu</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Fabio A. González</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Computing Systems and Industrial Engineering Dept., Universidad Nacional de Colombia Bogotá</institution>
          ,
          <country country="CO">Colombia</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Dept. of Computer Science, University of Houston Houston, TX</institution>
          ,
          <addr-line>77004</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Instituto Nacional de Astrofísica</institution>
          ,
          <addr-line>Óptica y Electrónica Puebla</addr-line>
          ,
          <country country="MX">Mexico</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2017</year>
      </pub-date>
      <abstract>
        <p>Social media data allows researchers to establish relationships between everyday language and people's sociodemographic variables, such as gender, age, language variety or personality. These variables configure social groups, where author profiling attempts to exploit the idea that they share a common language. This work describes our proposed method for the PAN 2017 Author Profiling shared task. We trained separate models for gender and language variety using a Convolutional Neural Network (CNN). We explored parameters such as the size of the input of the network, the size of the convolutional kernels, the number of kernels and the type of input. We found experimentally that sequences of words performed better than sequences of characters as input for the CNN. We obtained 0:66; 0:73; 0:81 and 0:57 of accuracy in the test partition for English, Spanish, Portuguese and Arabic respectively.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1 Introduction</title>
      <p>
        Author profiling consists of determining a social group of an unknown author [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
Chambers et al. support this idea with a sociolinguistics observation, where a social
group shares a way of speaking and writing, a dialect [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Several profile dimensions
for characterizing a social group have been considered since then, such as age, gender,
native language and personality. The relevance of this task has been recognized for its
applications that include forensics, marketing and security concerns.
      </p>
      <p>
        Last year’s PAN Author Profiling (PAN AP) shared task consisted of a cross-genre
age and gender prediction task [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. 22 teams participated using documents in English,
Spanish and Dutch. The teams approached this task using two kinds of features,
stylebased features and content-based features. Style-based features included n-gram
frequencies, punctuations, readability. Whereas content-based features comprise bag of
words, word n-grams, term vectors, named entities, among others. Previous successful
approaches have used style and content features. Argamon et al. showed
experimentally that content features performed better for language, age and gender profiling [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
However these features can be very sparse. López-Monroy et al. approached the
profiling task using a low-dimensional non-sparse representation of the documents of every
author [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. Other studies even describe that not all words matter when establishing the
profile of an author, but suggest that words near a personal pronoun are more
discriminative for classifying an author’s profile [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ].
      </p>
      <p>
        This year’s author profiling shared task consisted of predicting gender and
language variety of a group of authors in Twitter [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. This group of Twitter users is
distributed along four languages (English, Spanish, Portuguese and Arabic) and two
genders (Male and Female). Furthermore, language variety depended on the
respective language. English consisted of Australia, Canada, Great Britain, Ireland, New
Zealand and United States. Spanish consisted of Argentina, Chile, Colombia, Mexico,
Peru, Spain and Venezuela. Portuguese comprised Brazil and Portugal. Arabic included
Egypt, Gulf, Levantine and Maghrebi. Unlike PAN AP 2016 shared task [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], PAN AP
2017 shared task uses the same domain of documents (Twitter) for training and testing
[
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
      </p>
      <p>
        Also, this is the first time language variety identification is added to the PAN AP
shared task. Although language variety has been previously addressed as an author
profiling task [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. Furthermore, it has been part of the VarDial2016 (Workshop on NLP
for Similar Languages, Varieties and Dialects) [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Malmasi et al. describe a big gap
between traditional machine learning models and deep learning models in the participant
teams evaluated in the VarDial2016. In this work we attempt to narrow this gap using
convolutional neural networks (CNN) as a first approach for author profiling. CNNs
have proven to be a successful method for classification of texts [
        <xref ref-type="bibr" rid="ref5 ref6">6,5,17</xref>
        ]. CNNs have
also shown a good performance on authorship attribution tasks [
        <xref ref-type="bibr" rid="ref15">16,15</xref>
        ].
      </p>
      <p>In this paper we describe our approach using CNNs for the author profiling task.
CNNs are capable of capturing local-level interactions for learning profile-specific
patterns. For every language, we trained separate models for gender and language variety
using a CNN. Different CNN architectures were explored modifying parameters such
as the size of input of the network, the size of the convolutional kernels, the number
of kernels and the type of the input. We found experimentally that sequences of words
performed better than sequences of characters as input for the CNN.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Methodology</title>
      <p>
        In this section, we describe our submission to the PAN 2017 Author Profiling shared
task. This architecture is an adaptation of previous work of CNNs at character level [16].
First, we briefly describe the preprocessing strategy, then how CNNs work, and finally
we describe how CNN training was carried out. Authors’ tweets were tokenized using
a plain word tokenizer. We preserved both case and stopwords during preprocessing.
After that all the tweets of an author are concatenated and split into k evenly sized
sequences of texts. Words are then represented by non-sparse vectors of dimension e,
also known as embeddings. As Figure 1 shows, a sequence of words is represented as a
matrix C 2 Re k where each column corresponds to the word embedding vector value.
Word Convolutional Neural Networks (W-CNN) receive a fixed-length sequence of
words as input. Figure 1 depicts the W-CNN architecture. W-CNN first layer applies
a set of convolutional filters of different sizes. For the concrete case of Figure 1 m =
f500; 500; 500g and w = f2; 3; 4g. The convolution operation performed by these
filters is only applied in one dimension. Then a max-pooling over time operation is
performed over the output feature maps, where only the maximum value of each feature
map is used. The max pooling outputs for each feature map are concatenated in a
vector. Figure 1 shows the output vector of size 1500 composed by the maximum activation
values generated by each convolutional filter over the input. Finally, a softmax layer is
added, where its size An depends on the profiling task. Dropout regularization was also
used after the Embedding layer with a p = 0:25. Given that we train our network using
sequences of text of one author, we used a bagging scheme for prediction stage. If we
have n sequences of text for one author, we generate n predictions for the
corresponding author, then we average the predictions and get the class with the highest value. In
that way an author is labeled with its respective gender and language variety.
2.2 Implementation details
Several CNN architectures were explored for finding the most suitable models for the
author profiling task. Our exploration focused on two kinds of hyperparameters,
Inputrelated and Convolution-related. For Input-related parameters we explored the type of
input, the size of the input and the initialization values of the embeddings. The type
of input were either tokenized sequences of words or sequences of char bigrams. The
size of the input also was explored from a set of possible values f50; 100; 200; 300g.
Larger input sizes mean a reduction in the number of training samples, making the
training process difficult for complex architectures. Initialization values of the embeddings
were also evaluated using either pretrained embeddings or embeddings trained from
scratch using the supervised signal of the profiling task. Pretrained word embeddings
were trained on Wikipedia for every language using FastText [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], although, we found
that pretraining of the embeddings did not improve the results. For convolution-related
parameters we explored the size w of the kernels and the number of kernels m. Larger
size of kernels implies capturing long distance relationships between words, however
this is only possible with a sufficient amount of training samples. Accordingly, we
explored w from the set of values f1; 2; 3g; f2; 3; 4g; f4; 5; 6g, while the number of filters
m varied from 1500 up to 3000. Also using a large number of filters, increases the
representational capacity of the architecture, however it overfits quickly.
      </p>
      <p>
        These architecture hyperparameters were found by exploration on the validation
split of each setup and the best combination of parameters can be found in Table 1.
We found also that word-based inputs performed better than char-based inputs over all
the profiling setups. For training, we employed Keras [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. We shuffled the samples into
mini-batches of size 32 and used Gradient Descent with Adaptive Moment Estimation
[
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] with default learning rate. Validation loss was monitored during 100 epochs and
only models with the best validation accuracy were saved and used for testing.
      </p>
      <p>Layer Parameters GenEdenrglVisahriety GenSdperanVisahriety GePnodreturgVuaersieety GendAerrabViacriety</p>
      <p>Input type word word word word word word word word
Input Input size 200 200 300 200 200 200 50 300</p>
      <p>Pre-trained No No No No No No No No</p>
      <p>Convolutional mw [11;520;03] [11;520;03] [11;520;03] [11;520;03] [11;520;03] [11;520;03] [11;520;03] [11;520;03]
Table 1. Best combination of hyperparameters for the neural network architecture. Possible
values for the hyperparameters are as follows: Input type can be word or char. Input size varied from
f50; 100; 200; 300g. Pre-trained defined if embeddings were trained previously from Word2Vec
or not. Convolutional number of filters m varied from f1500; 3000g. Convolutional sizes of
filters w comprised f1; 2; 3g; f2; 3; 4g; f4; 5; 6g
This year’s training data consists of 10800 Twitter users. For each individual author, an
XML document is provided along with his/her tweets. There are 3000 documents for
English, 4200 for Spanish, 1200 for Portuguese and 2400 for Arabic. For each language,
we trained separately a model for gender and for language variety. For evaluation, we
generated a stratified train/val split for every possible combination of language_gender
and language_variety. Ten percent of the training documents was used for validation
purposes.</p>
      <p>
        The evaluation of the models for the shared task was performed using TIRA [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
TIRA allows both organizers and participants to have a common framework for
evaluation. Also, participants of a shared task can deploy and evaluate their method without
accessing directly to the test dataset. We deployed on TIRA the best model found by
validation. Table 2 shows the performance results of our method in the test dataset for
the four languages. Accuracy is calculated separately for gender and language variation.
For the joint column, accuracy is calculated on the basis that both gender and language
variation were properly predicted.
Our architecture was evaluated over sequences of words and characters. We found
experimentally better validation performances using word sequences. Although we also
found that training a CNN for author profiling produces additional challenges such as
hyperparameter tuning and quick overfitting. In our parameter exploration we
encountered models that were prone to overfit at the very first epochs. We solved this
introducing dropout regularization or using an architecture with a fewer number of parameters.
      </p>
      <p>
        This work is a first approximation to the author profiling task using neural networks.
Our system is capable of learning significant patterns without any handcrafted features,
however it still performs worse than traditional methods that use a concatenation of
content and style handcrafted features. Also, as it has been reported in previous works,
content-based representations work better than style-based. In our future work, we will
explore deeper convolutional networks with strong regularization, attention models and
oversampling strategies. Also as suggested in [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], feeding sequences of text centered
on personal pronouns could improve the performance of the CNN, because the network
would only look at relevant examples.
16. Shrestha, P., Sierra, S., Gonzalez, F., Montes, M., Rosso, P., Solorio, T.: Convolutional
neural networks for authorship attribution of short texts. In: Proceedings of the 15th
Conference of the European Chapter of the Association for Computational Linguistics:
Volume 2, Short Papers. pp. 669–674. Association for Computational Linguistics, Valencia,
Spain (April 2017), http://www.aclweb.org/anthology/E17-2106
17. Zhang, X., Zhao, J., LeCun, Y.: Character-level convolutional networks for text
classification. In: Advances in Neural Information Processing Systems. pp. 649–657 (2015)
      </p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Argamon</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Koppel</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pennebaker</surname>
            ,
            <given-names>J.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schler</surname>
          </string-name>
          , J.:
          <article-title>Automatically profiling the author of an anonymous text</article-title>
          .
          <source>Commun. ACM</source>
          <volume>52</volume>
          (
          <issue>2</issue>
          ),
          <fpage>119</fpage>
          -
          <lpage>123</lpage>
          (
          <year>Feb 2009</year>
          ), http://doi.acm.
          <source>org/10</source>
          .1145/1461928.1461959
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Bojanowski</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grave</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Joulin</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Enriching word vectors with subword information</article-title>
          .
          <source>arXiv preprint arXiv:1607.04606</source>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Chambers</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Trudgill</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schilling-Estes</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          :
          <article-title>The Handbook of Language Variation and Change</article-title>
          . Blackwell Handbooks in Linguistics, Wiley (
          <year>2002</year>
          ), https://books.google.com.co/books?id=1ImVXjkmbHkC
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Chollet</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          , et al.: Keras. https://github.com/fchollet/keras (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Kalchbrenner</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grefenstette</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Blunsom</surname>
            ,
            <given-names>P.:</given-names>
          </string-name>
          <article-title>A convolutional neural network for modelling sentences</article-title>
          .
          <source>In: Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)</source>
          . pp.
          <fpage>655</fpage>
          -
          <lpage>665</lpage>
          . Association for Computational Linguistics, Baltimore, Maryland (
          <year>June 2014</year>
          ), http://www.aclweb.org/anthology/P14-1062
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>Convolutional neural networks for sentence classification</article-title>
          .
          <source>In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP)</source>
          . pp.
          <fpage>1746</fpage>
          -
          <lpage>1751</lpage>
          . Association for Computational Linguistics, Doha, Qatar (
          <year>October 2014</year>
          ), http://www.aclweb.org/anthology/D14-1181
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Kingma</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ba</surname>
          </string-name>
          , J.:
          <article-title>Adam: a method for stochastic optimization</article-title>
          .
          <source>In: International Conference on Learning Representations</source>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>López-Monroy</surname>
            ,
            <given-names>A.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Montes-y Gómez</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Escalante</surname>
            ,
            <given-names>H.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Villaseñor-Pineda</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stamatatos</surname>
          </string-name>
          , E.:
          <article-title>Discriminative subprofile-specific representations for author profiling in social media</article-title>
          .
          <source>Knowledge-Based Systems 89</source>
          ,
          <fpage>134</fpage>
          -
          <lpage>147</lpage>
          (
          <year>2015</year>
          ), http://www.sciencedirect.com/science/article/pii/S0950705115002427
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Malmasi</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zampieri</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , Ljubešic´,
          <string-name>
            <given-names>N.</given-names>
            ,
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Ali</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Tiedemann</surname>
          </string-name>
          , J.:
          <article-title>Discriminating between similar languages and arabic dialect identification: A report on the third dsl shared task</article-title>
          .
          <source>In: Proceedings of the Third Workshop on NLP for Similar Languages, Varieties and Dialects (VarDial3)</source>
          . pp.
          <fpage>1</fpage>
          -
          <lpage>14</lpage>
          .
          <article-title>The COLING 2016 Organizing Committee</article-title>
          , Osaka,
          <source>Japan (December</source>
          <year>2016</year>
          ), http://aclweb.org/anthology/W16-4801
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Ortega-Mendoza</surname>
            ,
            <given-names>R.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Franco-Arcega</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>López-Monroy</surname>
            ,
            <given-names>A.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Montes-y Gómez</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>I, me, mine: The role of personal phrases in author profiling</article-title>
          .
          <source>In: International Conference of the Cross-Language Evaluation Forum for European Languages</source>
          . pp.
          <fpage>110</fpage>
          -
          <lpage>122</lpage>
          . Springer International Publishing (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Potthast</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gollub</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rangel</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stamatatos</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stein</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Improving the Reproducibility of PAN's Shared Tasks: Plagiarism Detection, Author Identification, and Author Profiling</article-title>
          . In: Kanoulas,
          <string-name>
            <given-names>E.</given-names>
            ,
            <surname>Lupu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Clough</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Sanderson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Hall</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Hanbury</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Toms</surname>
          </string-name>
          , E. (eds.)
          <article-title>Information Access Evaluation meets Multilinguality, Multimodality, and Visualization</article-title>
          .
          <source>5th International Conference of the CLEF Initiative (CLEF 14)</source>
          . pp.
          <fpage>268</fpage>
          -
          <lpage>299</lpage>
          . Springer, Berlin Heidelberg New York (
          <year>Sep 2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Rangel</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Franco-Salvador</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>P.:</given-names>
          </string-name>
          <article-title>A low dimensionality representation for language variety identification</article-title>
          .
          <source>In: Proceedings of the 17th International Conference on Computational Linguistics and Intelligent Text Processing (CICLing</source>
          <year>2016</year>
          ). Springer-Verlag (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Rangel</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Potthast</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stein</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Overview of the 5th Author Profiling Task at PAN 2017: Gender and Language Variety Identification in Twitter</article-title>
          . In: Cappellato,
          <string-name>
            <given-names>L.</given-names>
            ,
            <surname>Ferro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            ,
            <surname>Goeuriot</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            ,
            <surname>Mandl</surname>
          </string-name>
          , T. (eds.)
          <article-title>Working Notes Papers of the CLEF 2017 Evaluation Labs</article-title>
          .
          <source>CEUR Workshop Proceedings, CLEF and CEUR-WS.org (Sep</source>
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <given-names>Rangel</given-names>
            <surname>Pardo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            ,
            <surname>Rosso</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Verhoeven</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            ,
            <surname>Daelemans</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            ,
            <surname>Potthast</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Stein</surname>
          </string-name>
          ,
          <string-name>
            <surname>B.</surname>
          </string-name>
          :
          <article-title>Overview of the 4th Author Profiling Task at PAN 2016: Cross-Genre Evaluations</article-title>
          .
          <source>In: Working Notes Papers of the CLEF 2016 Evaluation Labs. CEUR Workshop Proceedings, CLEF and CEUR-WS.org (Sep</source>
          <year>2016</year>
          ), http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>1609</volume>
          /
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Ruder</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ghaffari</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Breslin</surname>
            ,
            <given-names>J.G.</given-names>
          </string-name>
          :
          <article-title>Character-level and multi-channel convolutional neural networks for large-scale authorship attribution</article-title>
          .
          <source>arXiv preprint arXiv:1609.06686</source>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>