<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Using Hashtags and POS-Tags for Author Profiling</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>ZHAW Zurich University of Applied Sciences</institution>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2019</year>
      </pub-date>
      <abstract>
        <p>This paper investigates automatic methods to separate human created from bot created Tweets, and in the case of human, determine the gender of an author. The novel contribution is the investigation of 2 research questions, firstly the usability of part of speech tags and secondly the usability of hashtags as additional features. It therefore extends the models presented by Daneshvar et al. and Basile et al. in the course of past Author Profiling Tasks @ Pan. The results are evaluated as part of the Author Profiling Task @ Pan 2019. It will be shown that the segmentation of hashtags as well as using POS-Tags n-grams can increase the accuracy when classifying bot and gender on the PAN Twitter-dataset. By adding these features and combining them in an ensemble classifier, it was possible to achieve accuracies of 94% for bots and 84% for gender for the English language on the official test set. However, with 79% for bots and 71% for gender, the performance on the Spanish part of the dataset differs significantly. Possible reasons for this shall be examined in the evaluation of the system.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>The ubiquity of social media, in private communication and media coverage calls for
strategies to validate both identity of users as well as the validity of the shared content
to prevent misuse and manipulation of public opinion.</p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] Shao et al. state that the deliberate spreading of false information, so-called
fake news, is a serious concern. Guess, Nagler and Tucker, who conducted a
representative online survey on Facebook users behavior in connection with fake news, say that
"The vast majority of Facebook users in our data did not share any articles from fake
news domains" [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. If one compares this with the finding from Chu, Gianvecchio, Wang
and Jajodia in [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] that 24% of the Tweets generated on Twitter originate from bots, and
relates it with the statement of Shao et al. that "social bots played a disproportionate
role in spreading articles from low-credibility sources" [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], it can be concluded that the
identification of bot profiles on social media is an important and promising approach to
prevent the spreading of fake news, and thus the manipulation of public opinion.
      </p>
      <p>The present work deals with the identification of features, which are suitable to
improve the accuracy of existing methods. The focus lies on the identification of bots
as well as the identification of gender from authors on Twitter. Both POS-tags as well
as the information contained in hashtags are considered.</p>
      <sec id="sec-1-1">
        <title>Research Questions</title>
        <sec id="sec-1-1-1">
          <title>The present work examines two central questions, namely:</title>
          <p>1. "Does the syntactic structure of Twitter Tweets reveal information about the
author’s identity with respect to gender/bot, moreover, if so, are such patterns
universal, i.e. are these patterns independent of content ?"
Part of speech tags were chosen to represent the syntactic structure. To consider the
sequential nature of the data, POS tags bi- and tri-grams were used as features..
2. "Do hashtags in Twitter Tweets contain information about the identity of the author
and can this information improve the accuracy of gender/bot-classification when
looking at the individual words they comprise?"
Due to the special nature of hashtags, where users are forced to use a single word,
therefore using compound words, the approach of segmenting hashtags and using
the resulting words as features was chosen.</p>
          <p>
            The goal is to develop a model that can classify Tweets via the use of an enriched
body of features, analyzing them in parallel and combining them.
1.2
In [
            <xref ref-type="bibr" rid="ref4">4</xref>
            ] author profiling is described as: "the analysis of shared content in order to
predict different attributes of authors such as gender, age, personality, native language, or
political orientation."
          </p>
          <p>
            The task described by Rangel et al. in [
            <xref ref-type="bibr" rid="ref5">5</xref>
            ] is concerned with the identification of an
authors gender and additionally if the author is either human or bot.
          </p>
          <p>
            PAN is a series of workshops concerned with digital text forensics [
            <xref ref-type="bibr" rid="ref6">6</xref>
            ] and is carried
out as part of the CLEF conference which is concerned with the systematic evaluation
of information access systems.
          </p>
          <p>The common basis of all participants is a dataset containing 100 Tweets per author,
which are combined in one file per author with a corresponding label assigned to it. A
label can have the following values: bot/human, and in the case of humans, female/male.</p>
          <p>The dataset of the Author Profiling Task 2019 includes the languages English and
Spanish. The English dataset contains 4120 authors, of which 2060 are bots, and the
remaining 2060 are divided into 1300 female and 1300 male authors. The Spanish dataset
contains 3000 authors, of which 1500 are bots, and the remaining 1500 are divided into
750 female and 750 male authors.</p>
          <p>
            Consequently, the gender of the author or whether the author is human should be
inferred based on a set of short messages which are available in purely textual form
(without meta information or additional content such as images). The evaluation of all
submitted systems is carried out on the online platform Tira [
            <xref ref-type="bibr" rid="ref7">7</xref>
            ].
          </p>
          <p>
            To obtain the final scores, the results of all participants are ranked by accuracy.
A detailed description as well as results and comparisons of systems submitted to the
Author Profiling Task @ Pan 2019 can be found in the official overview paper [
            <xref ref-type="bibr" rid="ref5">5</xref>
            ].
          </p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>
        A data set concerned with bot-detection is the honeypot data set. It was introduced by
Morstatter et al. in [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], and as the name suggests was created using so-called honeypot
bots. They say that honeypot bots such that "... any user in the network that connects to a
honeypot will be considered as a bot" [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. To identify bots, they developed an extension
of the AdaBoost algorithm which they call BoostOr, and used the honeypot dataset to
evaluate it. According to Morstatter et al. BoostOr "focuses more on the mislabeled bots
and downweights mis-labeled regular users." [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
      </p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] Cai, Li and Zengi Introduce their Behaviour enhanced deep bot detection
model. It is an artificial neural network architecture which uses an LSTM, to learn a
representation of the sequence of an authors Twitter history. They evaluated their model
on the honeypot dataset presented in [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], and report in [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] that this model, called BeDM
reaches an F1 score of 87.32% as opposed to the BoostOr model presented in [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] which,
according to Cai et al. reaches an F1 score of 86.10%
      </p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] Basile et al. presented a model with a "simple SVM system (using the
scikit-learn LinearSVM implementation) that uses character 3- to 5-grams and word 1- to
2-grams with tf-idf weighting" through which they achieved the best result in the Author
Profiling Task @ Pan 2017. In the following year, several of the best-ranked systems
(when only considering textual features) employed similar strategies with respect to
n-grams and the classification algorithm used.
      </p>
      <p>
        According to [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] the best result in the combined Author Profiling Task @ Pan 2018
was achieved by Takahashi et al. [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] Their text component consists of a bi-directional
recurrent neural network whose output leads via two successive pooling layers into a
fully connected layer. As features, they used word vectors. In this system, however, the
result of the textual features is supplemented with information from images, which are
analyzed using a convolutional neural network. The achieved accuracy averaged over
all three languages was according to [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] 78.72%.
      </p>
      <p>
        Rangel et al. state in [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] that the system of Daneshvar and Inkpen was able to achieve
the best results when only textual features were considered. The features were similar
to [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], with the addition of word 3-grams for the English part of the dataset and
subsequent Latent Semantic Analysis for all languages. The classification algorithm used
was a support vector machine [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ].
      </p>
      <p>
        According to [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] the accuracy averaged over all three languages for Daneshvar’s
model was 81.70%. This result is noteworthy as it shows that the best score of the 2018
task was achieved without the use of the provided images.
      </p>
      <p>
        According to [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] Tellez et al. achieved the second-best result when considering only
textual features, with a value of 80.99%. Similar to [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] and [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], the Bag of Words
approach was chosen using the tf-Idf weighting scheme, and support vector machines
for classification. Note, however, the additional use of skip grams [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
      </p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] Reuter, Pereira-Martins and Kalita present a pipeline to segment Hashtags.
It is a combination of several approaches, where the use of maximum known
matching seems to be worth mentioning, which tries to determine a metric for the length of
matches, and the result which delivers the longest match is rated highest.
      </p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] Declerck and Lendvai mention that Spanish sources contain fewer hashtags
than German ones, and that the camelCase notation is mainly used in English sources.
      </p>
      <p>Their approach segments hashtags written in camelCase notation in a first step, and
then uses them as a decision basis for segmenting hashtags written in lower case letters.</p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] and [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] it is stated that participants either normalized Tweets by removing
hashtags altogether, or used ratios of hashtags with respect to Tweets.
      </p>
      <p>To the author’s best knowledge, the approach of replacing hashtags with words
extracted by segmentation has not yet been used in a model submitted to the Pan
workshop.</p>
      <p>
        The use of POS tags as features in the form of n-grams has already been discussed
by Martinc et al. [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] in the Author Profiling Task @ Pan 2017, but only trigrams were
considered here. Furthermore, a single instance of a Logistic Regression Classifier was
employed for classification, to which a combination of differently weighted features
was provided.
      </p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] López-Santillán, Gonzalez-Gurrola and Ramfrez-Alonso introduce a model
in which they create embeddings of POS tags, using the same procedure as is used for
word embeddings. Here they chose the skip-gram approach. In addition to the word
embeddings the obtained document vector is then enriched with these POS-embeddings.
      </p>
      <p>
        The LDSE baseline by Rangel, Rosso and Franco is described under [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] and is
concerned with the Task of Language variety identification.
3
      </p>
    </sec>
    <sec id="sec-3">
      <title>Model Overview</title>
      <p>
        The model proposed in this paper is based on and extends the model presented by
Daneshvar et al. in [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] and Basile et al. in [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. It comprises two main components, of
which the first is the preprocessing pipeline, which is responsible for tokenization and
preprocessing of the Tweets.
      </p>
      <p>
        As is shown in Figure 1, the classification pipeline consists of a text and a POS-part,
of which the text component is similar to the implementation in [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], with
the addition of hashtag segmentation and handling of compound emojis. The second
component is concerned with the classification of POS-tags.
      </p>
      <p>The text and POS-part are combined in an ensemble classifier, which uses a support
vector machine as meta classifier.</p>
      <p>The resulting output is a prediction which is either bot/female/male, written to an
XML file per author.
This section provides a detailed discussion of the proposed model concerning the
research questions presented in the introduction. It first describes the preprocessing pipeline
and its specifics, with a particular focus on the peculiarities of Twitter Tweets, and then
the detailed structure of the classification pipeline.
As can be seen in Figure 2 during the preprocessing phase, the concatenated Tweets per
author are tokenized and Twitter specific replacements, hashtag-segmentation and part
of speech tagging is performed simultaneously.</p>
      <p>
        All of the preprocessing is performed using the spaCy NLP Framework as
introduced in [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ] by Honnibal and Montani. For the English part, the en_cor_web_sm
language model was used, which incorporates a convolutional neural network trained on
the OntoNotes corpus, consisting of blog articles, news, and comments [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ].
      </p>
      <p>
        For the Spanish part, the es_core_news_sm language model was used which is
trained on the AnCora and WikiNER corpus instead, which comprises news and media
content [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ]. The Twitter specific functionality was implemented via the extension of
custom pipeline objects provided by spaCy.
      </p>
      <sec id="sec-3-1">
        <title>Twitter Specific Preprocessing</title>
        <p>
          Substitutions
Similar to [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] the first step was to perform several substitutions on the concatenated
Tweets per author:
        </p>
        <sec id="sec-3-1-1">
          <title>Domain names: URLURL</title>
          <p>End of a tweet: SNTDLM
E-mail addresses: EMAILEMAIL
Twitter handles: USERMENTION</p>
          <p>Line breaks: LNFD</p>
          <p>
            To obtain accurate part of speech tags (POS tags), the replacement SNTDLM was
used to indicate the end of a sentence to the tagger explicitly. As with [
            <xref ref-type="bibr" rid="ref12">12</xref>
            ], sequences of
the same letters, which occurred more than three times, were replaced with a sequence
of 3 letters, resulting in a replacement of the following form:
          </p>
          <p>heeey, heeeeeey, heeeeeeey ! heeey
Custom POS tags
The jargon used in social media has some constructs that do not occur in verbal
communication or classical texts. In order to consider this, the POS tags have been enhanced
with the following elements:</p>
        </sec>
        <sec id="sec-3-1-2">
          <title>Domain names: URL</title>
          <p>Emojis: EMJI
E-mail addresses: EML
Hashtags: HSHT</p>
          <p>Twitter Handles: HNDL
Emojis
Emojis can be modified in several ways, e.g. there is a skin-tone modifier which can be
used to change the skin color of Emojis. Additionally, the combination of several Emojis
is possible, e.g. in the family Emoji , which consist of , and : Combining
several emojis is generally achieved by creating a sequence with the so-called
ZeroWidth-Joiner (ZWJ). When using a whitespace tokenizer, this is problematic in several
ways: firstly during tokenization, such sequences are cut at the ZWJ, and secondly,
the ZWJ remains in the resulting token stream. Therefore the tokenizer was adapted to
recognize compound emojis and treat them as one token. For the POS tagger this means
that regardless of the length of a sequence of emojis, the result is always one POS-tag.</p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>Hashtag Segmentation</title>
        <p>To determine whether hashtags contain information about the identity of an author,
when classifying bot/female/male, the procedure of segmenting composite hashtags
into individual words was chosen, resulting in replacements of the form:
#makeamericagreatagain ! make america great again
#roomforrent ! room for rent</p>
        <p>If the hashtag consist of a single word, a wordlist lookup is first performed to avoid
divisions such as the following:
#iconic ! i conic
#handsome ! hand some</p>
        <p>
          The Viterbi algorithm who was first presented by Viterbi in [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ] and more specific
an adaptation of it by Bacon [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ] was used to segment composite hashtags into
individual words. In [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ] it is described as follows: "the VA may be viewed as a solution
to the problem of maximum a posteriori probability (MAP) estimation of the state
sequence of a finite-state discrete-time Markov process". To calculate the probability of
the word under consideration, the algorithm needs to access word frequency lists.
During the development of the model, such lists were generated from the Pan Dataset, but
it was found that word frequency lists based on the OpenSubtitles corpus by Lison and
Tiedemann [
          <xref ref-type="bibr" rid="ref26">26</xref>
          ] gave superior results. Hence the final model uses them instead.
        </p>
        <p>In the actual algorithm a test is performed first, if the length of a hashtag is less than
3 characters, or it is contained within the provided wordlist, the word without the pound
character is returned.</p>
        <p>Then a nested loop is executed which steps through the string, considering each
substring contained in the hashtag, and using a function to assign it a probability.</p>
        <p>The mentioned function takes a word as argument and returns its probability, which
is calculated by dividing its frequency by the total number of word occurrences within
the provided word-frequency list (this information is contained data variable). The
words with the highest probability found are then returned.
4.2</p>
      </sec>
      <sec id="sec-3-3">
        <title>Classification</title>
        <p>
          The classification pipeline shown in Figure 3 consists of an ensemble that combines
the results of the text and POS components using an SVM meta-learner to make the
final prediction. The ensemble was implemented with the ML-Ensemble framework
developed by Flennerhag which facilitates parallel computations [
          <xref ref-type="bibr" rid="ref27">27</xref>
          ]. For the single
components such as the tf-idf vectorizer, singular value decomposition or the linear
svm classifier the sci-kit learn framework was used.
        </p>
        <p>Experiments were conducted with both, a soft- and hard-voting approach, it was
found that the ensemble achieves the best results using a soft-voting approach.</p>
      </sec>
      <sec id="sec-3-4">
        <title>N-Grams</title>
        <p>
          As proposed in [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] word 1- to 2-grams in addition to character 3- to 5-grams were
used in the text-component. As suggested in [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ], for English, character 3-grams were
also included. For the POS-Pipeline grid-search was performed which indicated that a
combination of word 2- and 3-grams are the optimal setting.
        </p>
      </sec>
      <sec id="sec-3-5">
        <title>Text Component</title>
        <p>
          The text component uses both word n-grams and character n-grams as features. Each of
which is transformed separately into tf-Idf vectors, where only tokens with a term
frequency greater than or equal to 2 are considered. The set of resulting document vectors
is the source material for latent semantic analysis. This part of the pipeline is essentially
an extension to the systems presented under [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]. Experiments with logistic
regression have been carried out. However, a support vector machine with a linear kernel has
proved to be the most effective choice, just like in [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ]. In order to enable multi-class
classification different strategies were considered out of which the best results were
achieved with a One vs. One approach.
        </p>
      </sec>
      <sec id="sec-3-6">
        <title>POS Component</title>
        <p>In the POS component, n-grams were also generated in a first step, but only on the
token level (no character n-grams were used). The use of n-grams was chosen to allow the
classifier to at least fundamentally analyze the information which lies in the sequence
of the data. Inspired by the text component, latent semantic analysis was also
experimented with but showed no improvement in accuracy. Interestingly, in contrast to the
text component, the accuracy increased when using logistic regression over a support
vector machine. Hence the final version uses it.</p>
      </sec>
      <sec id="sec-3-7">
        <title>Probability Calibration</title>
        <p>
          In [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ] Platt et al. explain that “Posterior probabilities are also required when a classifier
is making a small part of an overall decision, and the classification outputs must be
combined for the overall decision.” [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ]. He continues to point out that support vector
machines output an uncalibrated value which is not a probability. The ensemble of
the proposed model uses a meta-classifier with a soft-voting approach, which means
that it receives as input class-probabilities instead of hard labels. Therefore the
SVMcalssifier must be calibrated, as opposed to the logistic regression classifier of the
POScomponent, which according to Niculescu-Mizil and Caruan [
          <xref ref-type="bibr" rid="ref29">29</xref>
          ] already predicts
wellcalibrated probabilities. The calibration of the SVM was performed over the holdout set
of 3 folds; this step was directly included in the training process. In Figure 4 one can
see the reliability curve and the effect of calibration on the SVM-classifier.
5
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Evaluation</title>
      <p>This section presents and evaluates the results that the proposed model was able to
achieve on both the training and the test data. Special attention will be paid to the
performance of the POS-component and the differences between the Spanish and English
parts of the data-set.
5.1</p>
      <sec id="sec-4-1">
        <title>Results on the Training Data</title>
        <p>
          In accordance with [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] 60% of the PAN training data was used to train the models, and
40 % was used to evaluate them. In addition, 10-fold cross-validation was employed
during training. The following models were evaluated on the training data:
        </p>
        <sec id="sec-4-1-1">
          <title>1. Text component without hashtag-segmentation.</title>
          <p>2. Text component with hashtag-segmentation.
3. Ensemble with text-component and pos-component</p>
          <p>Looking at the results in Table 1, which lists the mean accuracies for each examined
model for English and Spanish, it is noticeable that although a wide range of features
has been considered, the differences are all below 2%. Nevertheless, the combination
of hashtag-segmentation and POS classification leads to a measurable improvement in
the overall result.</p>
          <p>Both hashtags and POS n-grams have a higher influence on the English part of the
data-set. The hashtags in the English part improve the accuracy from 92.6% to 94.4%,
whereas the accuracy in the Spanish part increases from 90.2% to 91.4%. The
differences for the POS component are even more pronounced. Looking at the differences
when adding the POS component, it can be seen that the increase in precision in the
English part is 0.5%, whereas in the Spanish part it is only 0.1%. Table 2 shows the
results with respect to precision for the final model on the training data by language and
class, here it shows, that the difference between class bot and gender for the Spanish
part is 5% higher than for the English part:</p>
          <p>Considered in isolation, with a mean accuracy of 81.3 % for Spanish, the POS
component has a significantly lower accuracy than the text component, but it improves
the overall result when used in an ensemble. The results for precision, recall, and
F1Score of the POS-component can be seen in Table 3</p>
          <p>Since a One vs. One approach was chosen to implement the multiclass-classification,
each of the three classes bot/female/male has its own instance per component. The
confusion matrix in Figure 5 now shows that in both languages the number of bots wrongly
classified as men was much higher than the number of bots classified as women. In
English 10 women compared to 15 men, and in Spanish even 4 women compared to 20
men.
5.2</p>
        </sec>
      </sec>
      <sec id="sec-4-2">
        <title>Results on the Test Data</title>
        <p>In order to evaluate the proposed model on the official test data set, it was first trained
on the entire training data and in a second step evaluated on the test data. The results
are listed in Table 4:</p>
        <p>What is particularly noteworthy about the results on the test data are the significant
differences between English and Spanish. The fact that the model achieved better results
on the English data-set could already be observed during training but was amplified on
the test data-set. Where on the training data-set the accuracy for Spanish was 84.5%
for the gender classification task compared to 89.5% for English, on the test data-set, it
became 71.2% for Spanish compared to 84% for English.
6</p>
        <p>Conclusion
1. Syntactic structure: It has been shown that it is possible to use POS-tags to
classify Tweets based on their syntactic structure, which means classification is possible
without any information about the actual content of a text. In addition, it was
determined during the evaluation that these features are suitable to improve the accuracy
of a system which until now has only classified on the basis of words.
2. Hashtags: The results presented as part of the evaluation show that it is possible
to improve accuracy when classifying, by segmenting the hashtags contained in the
Tweets into individual tokens/words and replacing them with the original hashtag.
It has also been shown that the approach presented, using word-frequency lists and
the Viterbi algorithm to perform this segmentation is feasible.</p>
        <p>
          However, it was not possible to determine what caused the large differences in
accuracy between Spanish and English. A possible explanation for this are the different
corpora used to train the Tokenizer/POS taggers in English and Spanish, and the
respective word-frequency lists. Although López-Santillán et al. do not use tf-idf vectors,
but embeddings of POS tags as features [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ], it is interesting that they also report lower
accuracies for Spanish than English.
7
        </p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Outlook</title>
      <p>It would be interesting to investigate to what extent longer sequences enable an
improvement in accuracy using an algorithm that is able to better address the
relationship between the individual elements. The use of LSTM or GRU networks would be
conceivable here. This would be a further step towards a model that can classify text
independent of content.</p>
      <p>In the proposed model, the tokens obtained by segmenting the hashtags were treated
exactly the same as other tokens. It should be examined whether a further improvement
in accuracy could be achieved through a different weighting scheme of the tokens
obtained from the hashtag segmentation.
8</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgements</title>
      <p>I would like to express my very great appreciation to Prof. Dr. Martin Braschler for his
valuable and constructive contribution to the planning and development of this paper.
His feedback as a supervisor has always been of great value to me.</p>
      <p>I would also like to thank Saman Daneshvar for providing the source code of his
model, which allowed me to focus on my research questions.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Shao</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ciampaglia</surname>
            ,
            <given-names>G.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Varol</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>K.C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Flammini</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Menczer</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>The spread of low-credibility content by social bots</article-title>
          .
          <source>Nature communications 9(1)</source>
          (
          <year>2018</year>
          )
          <fpage>4787</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Guess</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nagler</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tucker</surname>
          </string-name>
          , J.:
          <article-title>Less than you think: Prevalence and predictors of fake news dissemination on facebook</article-title>
          .
          <source>Science advances 5(1)</source>
          (
          <year>2019</year>
          ) eaau4586
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Chu</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gianvecchio</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jajodia</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Who is tweeting on twitter: human, bot, or cyborg? In: Proceedings of the 26th annual computer security applications conference</article-title>
          ,
          <source>ACM</source>
          (
          <year>2010</year>
          )
          <fpage>21</fpage>
          -
          <lpage>30</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Rangel</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Montes-y Gómez</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Potthast</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stein</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Overview of the 6th author profiling task at pan 2018: multimodal gender identification in twitter</article-title>
          .
          <source>In: CEUR Workshop Proceedings</source>
          , CEUR, CEUR-WS.org (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Rangel</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cappellato</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ferro</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Müller</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Losada</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Overview of the 7th author profiling task at pan 2019: Bots and gender profiling</article-title>
          . In Cappellato, L.,
          <string-name>
            <surname>Ferro</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Losada</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Müller</surname>
          </string-name>
          , H., eds.
          <source>: CEUR Workshop Proceedings</source>
          , CEUR, CEUR-WS.
          <source>org (September</source>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Daelemans</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kestemont</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manjavancas</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Potthast</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rangel</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Specht</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stamatatos</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stein</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tschuggnall</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wiegmann</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zangerle</surname>
          </string-name>
          , E.: Overview of PAN 2019:
          <article-title>Author Profiling, Celebrity Profiling, Cross-domain Authorship Attribution and Style Change Detection</article-title>
          . In
          <string-name>
            <surname>Crestani</surname>
          </string-name>
          , F.,
          <string-name>
            <surname>Braschler</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Savoy</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rauber</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Müller</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Losada</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Heinatz</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cappellato</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ferro</surname>
          </string-name>
          , N., eds.
          <source>: Proceedings of the Tenth International Conference of the CLEF Association (CLEF</source>
          <year>2019</year>
          ), Springer (
          <year>September 2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Potthast</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gollub</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wiegmann</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stein</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>TIRA Integrated Research Architecture</article-title>
          . In Ferro, N.,
          <string-name>
            <surname>Peters</surname>
          </string-name>
          , C., eds.
          <source>: Information Retrieval Evaluation in a Changing World - Lessons Learned from 20 Years of CLEF</source>
          . Springer (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Morstatter</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nazer</surname>
            ,
            <given-names>T.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Carley</surname>
            ,
            <given-names>K.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liu</surname>
          </string-name>
          , H.:
          <article-title>A new approach to bot detection: striking the balance between precision and recall</article-title>
          .
          <source>In: 2016 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM)</source>
          ,
          <source>IEEE</source>
          (
          <year>2016</year>
          )
          <fpage>533</fpage>
          -
          <lpage>540</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Cai</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zengi</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Behavior enhanced deep bot detection in social media</article-title>
          .
          <source>In: 2017 IEEE International Conference on Intelligence and Security Informatics (ISI)</source>
          ,
          <source>IEEE</source>
          (
          <year>2017</year>
          )
          <fpage>128</fpage>
          -
          <lpage>130</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Basile</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dwyer</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Medvedeva</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rawee</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Haagsma</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nissim</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>N-gram: New groningen author-profiling model</article-title>
          .
          <source>arXiv preprint arXiv:1707.03764</source>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Takahashi</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tahara</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nagatani</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Miura</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Taniguchi</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ohkuma</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Text and image synergy with feature cross technique for gender identification</article-title>
          .
          <source>Working Notes Papers of the CLEF</source>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Daneshvar</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Inkpen</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Gender identification in twitter using n-grams and lsa</article-title>
          .
          <source>In: Proceedings of the Ninth International Conference of the CLEF Association (CLEF</source>
          <year>2018</year>
          ).
          <article-title>(</article-title>
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Tellez</surname>
            ,
            <given-names>E.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Miranda-Jiménez</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moctezuma</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Graff</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Salgado</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ortiz-Bejar</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          :
          <article-title>Gender identification through multi-modal tweet analysis using microtc and bag of visual words</article-title>
          .
          <source>In: Proceedings of the Ninth International Conference of the CLEF Association (CLEF</source>
          <year>2018</year>
          ).
          <article-title>(</article-title>
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Reuter</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pereira-Martins</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kalita</surname>
          </string-name>
          , J.:
          <article-title>Segmenting twitter hashtags</article-title>
          .
          <source>Intl. J. on Natural Lang. Computing</source>
          <volume>5</volume>
          (
          <issue>4</issue>
          ) (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Declerck</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lendvai</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Processing and normalizing hashtags</article-title>
          .
          <source>In: Proceedings of the International Conference Recent Advances in Natural Language Processing</source>
          . (
          <year>2015</year>
          )
          <fpage>104</fpage>
          -
          <lpage>109</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Rangel</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Potthast</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stein</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Overview of the 5th author profiling task at pan 2017: Gender and language variety identification in twitter</article-title>
          .
          <source>In: Working Notes Papers of the CLEF, CEUR, CEUR-WS.org</source>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Martinc</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Skrjanec</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zupan</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pollak</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          : Pan 2017:
          <article-title>Author profiling-gender and language variety prediction</article-title>
          .
          <source>In: CLEF (Working Notes)</source>
          . (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>López-Santillán</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gonzalez-Gurrola</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ramfrez-Alonso</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          :
          <article-title>Custom document embeddings via the centroids method: Gender classification in an author profiling task</article-title>
          .
          <source>In: Proceedings of the Ninth International Conference of the CLEF Association (CLEF</source>
          <year>2018</year>
          ).
          <article-title>(</article-title>
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Rangel</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Franco-Salvador</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>P.:</given-names>
          </string-name>
          <article-title>A low dimensionality representation for language variety identification</article-title>
          .
          <source>In: Proceedings of the 17th International Conference on Intelligent Text Processing and Computational Linguistics (CICLing'16)</source>
          , Springer (
          <year>2018</year>
          )
          <fpage>156</fpage>
          -
          <lpage>169</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Honnibal</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Johnson</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>An improved non-monotonic transition system for dependency parsing</article-title>
          .
          <source>In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing</source>
          , Lisbon, Portugal, Association for Computational Linguistics (
          <year>September 2015</year>
          )
          <fpage>1373</fpage>
          -
          <lpage>1378</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Honnibal</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Montani</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>English · spacy models documentation https://spacy</article-title>
          .io/models/en.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Honnibal</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Montani</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Spanish · spacy models documentation https://spacy</article-title>
          .io/models/es.
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Viterbi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Error bounds for convolutional codes and an asymptotically optimum decoding algorithm</article-title>
          .
          <source>IEEE transactions on Information Theory</source>
          <volume>13</volume>
          (
          <issue>2</issue>
          ) (
          <year>1967</year>
          )
          <fpage>260</fpage>
          -
          <lpage>269</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24.
          <string-name>
            <surname>Bacon</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>How can i split multiple joined words</article-title>
          ? http://stackoverflow.com/a/481773/554406.
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          25.
          <string-name>
            <surname>Forney</surname>
            ,
            <given-names>G.D.:</given-names>
          </string-name>
          <article-title>The viterbi algorithm</article-title>
          .
          <source>Proceedings of the IEEE 61(3)</source>
          (
          <year>1973</year>
          )
          <fpage>268</fpage>
          -
          <lpage>278</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          26.
          <string-name>
            <surname>Lison</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tiedemann</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>Opensubtitles2016: Extracting large parallel corpora from movie and tv subtitles</article-title>
          . (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          27.
          <string-name>
            <surname>Flennerhag</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          : Ml-ensemble http://ml-ensemble.com/.
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          28.
          <string-name>
            <surname>Platt</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          , et al.:
          <article-title>Probabilistic outputs for support vector machines and comparisons to regularized likelihood methods</article-title>
          .
          <source>Advances in large margin classifiers 10(3)</source>
          (
          <year>1999</year>
          )
          <fpage>61</fpage>
          -
          <lpage>74</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          29.
          <string-name>
            <surname>Niculescu-Mizil</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Caruana</surname>
          </string-name>
          , R.:
          <article-title>Predicting good probabilities with supervised learning</article-title>
          .
          <source>In: Proceedings of the 22nd international conference on Machine learning</source>
          ,
          <source>ACM</source>
          (
          <year>2005</year>
          )
          <fpage>625</fpage>
          -
          <lpage>632</lpage>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>