<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>SU@PAN'2015: Experiments in Author Pro ling</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Yasen Kiprov</string-name>
          <email>yasen.kiprov@gmail.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Momchil Hardalov</string-name>
          <email>momchil.hardalov@gmail.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Preslav Nakov</string-name>
          <email>pnakov@qf.org.qa</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ivan Koychev</string-name>
          <email>koychev@fmi.uni-sofia.bg</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Qatar Computing Research Institute</institution>
          ,
          <addr-line>HBKU</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>So a University \St. Kliment Ohridski"</institution>
          ,
          <country country="BG">Bulgaria</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>We describe the submission of the So a University team for the Author Pro ling Task, part of the PAN 2015 Challenge. Given a set of writing samples by the same person, the task asks to predict some demographical information such as age and gender, as well as the personality type of that person. We experimented with SVM classi ers using variety of features extracted from publicly available resources, achieving the second-best score for Spanish out of 21 submissions, and the sixthbest for English out of 22 submissions.</p>
      </abstract>
      <kwd-group>
        <kwd>author pro ling</kwd>
        <kwd>text mining</kwd>
        <kwd>machine learning</kwd>
        <kwd>PAN 2015</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>Social media applications such as Facebook and Twitter have hundreds of
millions of users who share information at an unprecedented scale. Naturally, this
has attracted business and research interest from various elds including
marketing, political science, and social studies, among others. Initially, the primary
research and practical interest was in the opinion and the sentiment users
express, e.g., towards products such as iPhone6, policies such as ObamaCare, and
events such as Pope's visit to Palestine.</p>
      <p>Soon, companies realized that knowing the overall public opinion towards
their products was not enough; for marketing purposes, it was also important
to be able to break this opinion by demographic factors such as gender and age.
Similarly, public policy and security experts wanted to further know the native
language and some personality traits of the users expressing particular opinions.
The personality traits were also important for human resources experts when
considering to hire somebody in a team: they wanted to make sure the person's
personality would t well in the target team.</p>
      <p>While some demographic information was sometimes directly extractable
from the public user pro les, there was no guarantee it was correct and
upto-date, which motivated research in trying to predict it automatically from the
text of the messages posted by a given user.</p>
      <p>
        Early work on detecting sentiment focused on newswire text [
        <xref ref-type="bibr" rid="ref13 ref21 ref3 ref30">3, 13, 21, 30</xref>
        ].
As subsequently research turned towards social media, it became clear that this
presented a number of new challenges. Misspellings, poor grammatical
structure, emoticons, acronyms, and slang were common in these new media, and
were explored by a number of researchers [
        <xref ref-type="bibr" rid="ref11 ref14 ref15 ref19 ref20 ref4 ref5">4, 5, 11, 14, 15, 19, 20</xref>
        ]. Later,
specialized shared tasks emerged, e.g., at SemEval, the International Workshop on
Semantic Evaluation, [
        <xref ref-type="bibr" rid="ref18 ref27 ref28">18, 27, 28</xref>
        ], which ran in 2013{2015 and compared systems
by participating teams against each other in a controlled environment using the
same training and testing datasets.
      </p>
      <p>
        A similar research trend followed with respect to author pro ling. While
there were several publications that were trying to predict some
demographical information such as gender, age, and native language [
        <xref ref-type="bibr" rid="ref2 ref23">2, 23</xref>
        ], as well as the
personality type, and to perform author pro ling in general [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], the real push
forward was enabled by specialized competitions such as the PAN shared tasks
on author pro ling [
        <xref ref-type="bibr" rid="ref22 ref25 ref26">22, 26, 25</xref>
        ], which ran in 2013{2015.
      </p>
      <p>
        Below we discuss the participation of the So a University team, registered
as kiprov15, in the 2015 edition of the task, which ran as part of the PAN
2015 [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ], the 13th evaluation lab on uncovering plagiarism, authorship, and
social software misuse. The task focused on predicting an author's demographics
(age and gender) and the big ve personality traits [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] (agreeable, conscientious,
extroverted, open, stable) from the text of a set of tweets by the same target
author.
      </p>
      <p>The task was o ered in English, Spanish, Dutch and Italian, but we
participated with a system for the rst two languages only. We experimented with
SVM classi ers using variety of features extracted from publicly available
resources, achieving the second-best score for Spanish out of 21 submissions, and
the sixth-best for English out of 22 submissions.</p>
      <p>We built our system from scratch, as a M.Sc. class project; our research,
the lessons we learned and some observations about the topic and potentially
relevant references, datasets, and resources are presented in the next sections. We
should note that we have saved a lot of time by reusing the GATE infrastructure
and by performing feature extraction on top of it. We focused most of our e orts
on feature engineering: we implemented some previously-proposed features, and
we further analyzed the training data in an attempt to design some new ones.</p>
      <p>The remainder of this paper is organized as follows: Section 2 gives an
overview of our approach, including a description of the preprocessing, the
features, and the learning algorithm we used. Section 3 presents our experimental
setup and the o cial results our system achieved. Section 4 discusses our results
and provides some deeper analysis. Finally, Section 5 concludes and points to
possible directions for future work.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Method</title>
      <p>
        We use the GATE framework [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] with various plugins to extract features from
the input documents. We then use these features in an SVM classi er, which
is implemented as a GATE plugin and uses the LibSVM library [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. The whole
system is wrapped in a Java console application and can be executed from the
command line.
2.1
      </p>
      <sec id="sec-2-1">
        <title>Preprocessing</title>
        <p>
          We integrated a pipeline of various resources for text analysis that are already
available in GATE such as a Twitter-speci c tokenizer [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ], a regular
expressionbased sentence splitter, a language identi er inplemented as a GATE plugin
based on TextCat [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ], the OpenNLP5 POS tagger, and a Twitter POS Tagger
[
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] for English. We further implemented custom text processing components in
order to handle emotions, elongated words, and punctuation. Finally, we
integrated various dictionaries and lexicons to detect sentiment polarity, emotions,
profanity, and personality-speci c phrases in the tweets. We processed the
documents using a pipeline with the following components:
1. Twitter tokenizer
2. RegEx sentence splitter
3. Language identi er
4. Language-speci c feature extractors (POS tags)
5. Gazetteer lookups
6. Rule-based feature extractors
7. Classi ers
2.2
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>Features</title>
        <p>
          Lexicon Features. For English, we used several lexicons, both
manuallycrafted and automatically generated:
{ NRC Hashtag Emotion Lexicon [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]: 16,862 terms with emotion;
{ Bad words lexicon: a combination of Google's \what do you love"
profanity dictionary6 and a manually assembled dictionary, with 874 terms in
total;
{ World Well-Being Project Personality Lexicon [
          <xref ref-type="bibr" rid="ref29">29</xref>
          ]: top and bottom
100 words for each of the ve traits with their score, a total of 1000 words.
        </p>
        <p>For each lexicon, we found in the tweets the terms that were listed in it, and
then we added as features the total terms count, normalized by the total number
of tweets. We instantiated separate features for the di erent lexicons.</p>
        <p>For the Hashtag Emotion Lexicon, we also calculated the total score of all
matching terms for each emotion type: anticipation, fear, anger, trust, surprise,
sadness, joy, and disgust.</p>
        <p>For the World Well-Being Project Personality Lexicon, we calculated a
total of 30 features, i.e., the following six features for each of the big ve types
(extroverted, stable, agreeable, conscientious, and open):
5 https://opennlp.apache.org/
6 http://www.wdyl.com/
{ Total positive terms score and count;
{ Total negative terms score and count;
{ Total terms score and count.</p>
        <p>Twitter-speci c Features. We used the following Twitter-speci c features:
{ Letter case: the number of lower-case, all-caps, and mix-case words;
{ Hashtags: the number of hashtags;
{ URLs: the number of URLs posted;
{ Retweets: the number of retweets;
{ User mentions: the number of mentions of users using the pattern @username;
{ User mentions start: the number of tweets starting with a user mention;
{ Picture share: the number of shared pictures using the pattern [pic]. This
feature was eventually dropped because there were too few users sharing this
type of content, and thus using it did not yield improvements.</p>
        <p>
          All of the above counts are normalized by the total number of available tweets
for the target user; so they could be viewed as \average number per tweet".
Orthographic Features. We used the following orthographic features:
{ Elonged words: the number of words with a sequence of more than two
identical characters;
{ Average sentence length: the average length of a sentence.
Term-level Features. We used the following term-level features:
{ n-grams: presence and count of unigrams and bigrams. This feature helps
to nd similar users based on their vocabulary overlap. For tokenizing the
text, we used the GATE Twitter-speci c tokenizer, which is aware of URLs,
emoticons, Twitter tags, etc.
{ Vocabulary size: the number of di erent words used by a user in all his
tweets.
{ POS tagging: We used a specialized POS tagger for tweets in English,
TwitIE, which is available in GATE [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]; it uses the Penn Treebank tagset,
but is optimized for tweets. For Spanish, we used OpenNLP, with pre-trained
models again in the Penn Treebank tagset, but the results were not so
accurate because of the tweet speci cs. Using these toolkits, we performed POS
tagging for both languages, and we extracted all POS tag types used in the
tweet together with their frequencies as features.
2.3
        </p>
      </sec>
      <sec id="sec-2-3">
        <title>Classi cation/Regression</title>
        <p>We used the above features and support vector machines (SVM) as implemented
in LibSVM. We trained a separate classi er for each language and for each target
category: one for age, one for gender. As the ve personality traits (extroverted,
stable, agreeable, conscientious, and open) were real values, we predicted them
using support vector regression. For both classi cation and regression, we used
the features as they were generated in the feature-extraction phase without any
further scaling or normalization, as most of them were in reasonable ranges,
e.g., because they were already normalized by the number of tweets.</p>
        <p>Since the challenge contained documents from di erent languages and topics,
we aimed to avoid as much as possible the use of language-speci c markers. Note
that most of our features are token-based.</p>
        <p>For training, we used linear kernels, which are known to be su cient for text
classi cation: as we had a very large number of unigrams and bigrams, there was
no need to use a kernel in order to make the two classes linearly separable.
3</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Experiments and Evaluation</title>
      <p>3.1</p>
      <sec id="sec-3-1">
        <title>Experimental Setup</title>
        <p>During development, we trained on training datasets provided by the organizers:
{ pan15-author-pro ling-training-dataset-english-2015-03-02
{ pan15-author-pro ling-training-dataset-spanish-2015-03-02</p>
        <sec id="sec-3-1-1">
          <title>We then tested on the following datasets:</title>
          <p>{ pan15-author-pro ling-test-dataset2-english-2015-04-23
{ pan15-author-pro ling-test-dataset2-spanish-2015-04-23</p>
          <p>While developing the system, we randomly selected 15 percent of our training
data to try out new features and to measure progress. For the o cial submission,
we trained our model on all provided training data, and we tested on the test
datasets.
3.2</p>
          <p>O</p>
          <p>
            cial Results
We were ranked 6th out of 22 submissions for English, and 2nd out of 21
submissions for Spanish. A summary of our o cial results is shown in Table 1, where
we have included our GLOBAL and RMSE scores, as well as our ranking among
the other participants in PAN-2015 Author pro ling task[
            <xref ref-type="bibr" rid="ref25">25</xref>
            ]. Table 2 shows our
o cial scores for each of the individual categories: age, gender and each of the
ve pro le traits.
          </p>
          <p>Although we also experimented with Dutch and Italian, we did not have
enough time to ne-tune our systems for them, and thus we eventually decided
not to submit results for them.
In our experiments above, we tried to solve the author pro ling task as a standard
n-gram based classi cation/regression problem, treating the target classes as
nominal (age, gender) or as real-valued (the big ve traits).</p>
          <p>We started with the same types of features for both language, only adding
dictionaries for English. Thus, our general method is in principle
languageindependent and applicable to any language for which data is available.</p>
          <p>We further did a lot of extra feature engineering and selection for English,
but we did not have time to do anything special in that respect for Spanish. Yet,
we performed better for Spanish, which was a surprise for us.</p>
          <p>
            Over the development of our nal model, we tested various combinations of
features, and we eventually excluded groups of features that did not actually
improve the performance. One such feature group were n-grams with length
more than 2; our observation is that 3-grams and 4-grams, although helpful
for sentiment analysis [
            <xref ref-type="bibr" rid="ref17">17</xref>
            ] on corpora with comparable size, did not improve
the performance for author pro ling. It looks like the tweet discourse is less
important than the actual vocabulary.
          </p>
          <p>Most of the orthographic features and dictionaries did not improve signi
cantly the big ve score, although improving the age and gender accuracy.
Removing all dictionaries drops the average age and gender F1 score by 2.5 percent
while not changing any of the big ve RMSEs by more than 0.005.</p>
          <p>We further tested the assumption that advertising and posting titles and
news articles introduces noise regarding personality detection by removing all
upper-case words from the n-grams. However, this worsened the RMSE by 0.02
on average.</p>
          <p>
            We believe the LIWC [
            <xref ref-type="bibr" rid="ref24">24</xref>
            ] resources could improve the accuracy of our model.
However, we decided to throw extra e ort in expanding the training corpus with
unsupervised data, to the point where LIWC features would be covered by our
own, trained on enough data. It is easy to detect user personality indications on
Twitter as many people tweet results from personality tests online; however, our
investigation shows that most of them follow the Myers-Briggs indicator, which
does not correlate well with the big ve. Thus, we did not use any extra training
data as we managed to nd very few examples per language.
          </p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Conclusion and Future Work</title>
      <p>We have described the system built by So a University's kiprov15 team for the
PAN-2015 Author Pro ling task, which was ranked 2nd in Spanish and 6th in
English, according to the o cial GLOBAL score.</p>
      <p>We have made some interesting observations about the impact of the di erent
features. Among the best feature groups were POS-tag counts and word unigrams
and bigrams. These had the most sustainable performance over the provided test
datasets.</p>
      <p>Even though we managed to achieve good ranking for the two languages we
made submissions for, we feel that there is a lot more to gain. For example, we
would like to try using di erent word clusters, thesaurus, the Linguistic Inquiry
and Word Count (LIWC) lexicons for di erent languages, named entity
recognition and normalization, e.g., locations, dates, numbers, money, person names,
etc.</p>
      <p>In addition to adding extra features, we are interested in using the social
media to generate more training examples. In particular, we would like to explore
the way personality is expressed in Twitter and whether it is dependent upon
language usage in general. For instance, would the model improve if we have
di erent training sets for English-speaking users in USA vs. UK vs. Canada vs.
Australia vs. India, etc.
6</p>
    </sec>
    <sec id="sec-5">
      <title>Source Code</title>
      <sec id="sec-5-1">
        <title>The project source code can be found on GitHub: https://github.com/ykiprov/pan2015</title>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <given-names>Shlomo</given-names>
            <surname>Argamon</surname>
          </string-name>
          , Moshe Koppel, James W. Pennebaker, and
          <string-name>
            <given-names>Jonathan</given-names>
            <surname>Schler</surname>
          </string-name>
          .
          <article-title>Automatically pro ling the author of an anonymous text</article-title>
          .
          <source>Commun. ACM</source>
          ,
          <volume>52</volume>
          (
          <issue>2</issue>
          ):
          <volume>119</volume>
          {
          <fpage>123</fpage>
          ,
          <year>February 2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <given-names>Shlomo</given-names>
            <surname>Argamon</surname>
          </string-name>
          and
          <article-title>Anat Rachel Shimoni</article-title>
          .
          <article-title>Automatically categorizing written texts by author gender</article-title>
          .
          <source>Literary and Linguistic Computing</source>
          ,
          <volume>17</volume>
          :
          <fpage>401</fpage>
          {
          <fpage>412</fpage>
          ,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <given-names>Stefano</given-names>
            <surname>Baccianella</surname>
          </string-name>
          ,
          <source>Andrea Esuli, and Fabrizio Sebastiani. SentiWordNet 3</source>
          .
          <article-title>0: An enhanced lexical resource for sentiment analysis and opinion mining</article-title>
          .
          <source>In Proceedings of the Seventh International Conference on Language Resources and Evaluation</source>
          , LREC '
          <volume>10</volume>
          , pages
          <fpage>2200</fpage>
          {
          <fpage>2204</fpage>
          ,
          <string-name>
            <surname>Valletta</surname>
          </string-name>
          , Malta,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <given-names>Luciano</given-names>
            <surname>Barbosa</surname>
          </string-name>
          and
          <string-name>
            <given-names>Junlan</given-names>
            <surname>Feng</surname>
          </string-name>
          .
          <article-title>Robust sentiment detection on Twitter from biased and noisy data</article-title>
          .
          <source>In Proceedings of the 23rd International Conference on Computational Linguistics: Posters, COLING '10</source>
          , pages
          <fpage>36</fpage>
          {
          <fpage>44</fpage>
          , Beijing, China,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Albert</surname>
            <given-names>Bifet</given-names>
          </string-name>
          , Geo rey Holmes, Bernhard Pfahringer, and
          <string-name>
            <given-names>Ricard</given-names>
            <surname>Gavalda</surname>
          </string-name>
          .
          <article-title>Detecting sentiment change in Twitter streaming data</article-title>
          .
          <source>Journal of Machine Learning Research, Proceedings Track</source>
          ,
          <volume>17</volume>
          :5{
          <fpage>11</fpage>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <given-names>Kalina</given-names>
            <surname>Bontcheva</surname>
          </string-name>
          , Leon Derczynski, Adam Funk,
          <string-name>
            <given-names>Mark A.</given-names>
            <surname>Greenwood</surname>
          </string-name>
          , Diana Maynard, and Niraj Aswani.
          <article-title>TwitIE: An open-source information extraction pipeline for microblog text</article-title>
          .
          <source>In Proceedings of the International Conference on Recent Advances in Natural Language Processing, RANLP '13</source>
          , pages
          <fpage>83</fpage>
          {
          <fpage>90</fpage>
          ,
          <string-name>
            <surname>Hissar</surname>
          </string-name>
          , Bulgaria,
          <year>2013</year>
          . INCOMA Ltd.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>William</surname>
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Cavnar and John M. Trenkle</surname>
          </string-name>
          .
          <article-title>N-gram-based text categorization</article-title>
          .
          <source>In Proceedings of the 3rd Annual Symposium on Document Analysis and Information Retrieval</source>
          , SDAIR '
          <volume>94</volume>
          , pages
          <fpage>161</fpage>
          {
          <fpage>175</fpage>
          ,
          <string-name>
            <surname>Las</surname>
            <given-names>Vegas</given-names>
          </string-name>
          , Nevada,
          <year>1994</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Chih-Chung Chang</surname>
          </string-name>
          and
          <string-name>
            <surname>Chih-Jen Lin</surname>
          </string-name>
          .
          <article-title>LIBSVM: A library for support vector machines</article-title>
          .
          <source>ACM Transactions on Intelligent Systems and Technology</source>
          ,
          <volume>2</volume>
          :
          <issue>27</issue>
          :1{
          <fpage>27</fpage>
          :
          <fpage>27</fpage>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9. Paul T Costa and
          <string-name>
            <surname>Robert R McCrae.</surname>
          </string-name>
          <article-title>The revised neo personality inventory (neopi-r)</article-title>
          .
          <source>The SAGE handbook of personality theory and assessment</source>
          ,
          <volume>2</volume>
          :
          <fpage>179</fpage>
          {
          <fpage>198</fpage>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Hamish</surname>
            <given-names>Cunningham</given-names>
          </string-name>
          , Diana Maynard, Kalina Bontcheva, Valentin Tablan, Niraj Aswani, Ian Roberts, Genevieve Gorrell, Adam Funk, Angus Roberts, Danica Damljanovic, Thomas Heitz,
          <string-name>
            <given-names>Mark A.</given-names>
            <surname>Greenwood</surname>
          </string-name>
          , Horacio Saggion, Johann Petrak,
          <string-name>
            <given-names>Yaoyong</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>and Wim</given-names>
            <surname>Peters</surname>
          </string-name>
          .
          <source>Text Processing with GATE (Version 6)</source>
          .
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Dmitry</surname>
            <given-names>Davidov</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Oren</given-names>
            <surname>Tsur</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Ari</given-names>
            <surname>Rappoport</surname>
          </string-name>
          .
          <article-title>Semi-supervised recognition of sarcasm in Twitter and Amazon</article-title>
          .
          <source>In Proceedings of the Fourteenth Conference on Computational Natural Language Learning</source>
          ,
          <source>CoNLL '10</source>
          , pages
          <fpage>107</fpage>
          {
          <fpage>116</fpage>
          ,
          <string-name>
            <surname>Uppsala</surname>
          </string-name>
          , Sweden,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Leon</surname>
            <given-names>Derczynski</given-names>
          </string-name>
          , Alan Ritter, Sam Clark, and
          <string-name>
            <given-names>Kalina</given-names>
            <surname>Bontcheva</surname>
          </string-name>
          .
          <article-title>Twitter partof-speech tagging for all: Overcoming sparse and noisy data</article-title>
          .
          <source>In Proceedings of the International Conference on Recent Advances in Natural Language Processing, RANLP '13</source>
          , pages
          <fpage>198</fpage>
          {
          <fpage>206</fpage>
          ,
          <string-name>
            <surname>Hissar</surname>
          </string-name>
          , Bulgaria,
          <year>2013</year>
          . Incoma Ltd.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <given-names>Minqing</given-names>
            <surname>Hu</surname>
          </string-name>
          and
          <string-name>
            <given-names>Bing</given-names>
            <surname>Liu</surname>
          </string-name>
          .
          <article-title>Mining and summarizing customer reviews</article-title>
          .
          <source>In Proceedings of the 10th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD '04</source>
          , pages
          <fpage>168</fpage>
          {
          <fpage>177</fpage>
          , New York, NY, USA,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Bernard</surname>
            <given-names>Jansen</given-names>
          </string-name>
          , Mimi Zhang, Kate Sobel, and
          <string-name>
            <given-names>Abdur</given-names>
            <surname>Chowdury</surname>
          </string-name>
          .
          <article-title>Twitter power: Tweets as electronic word of mouth</article-title>
          .
          <source>J. Am. Soc. Inf. Sci. Technol</source>
          .,
          <volume>60</volume>
          (
          <issue>11</issue>
          ):
          <volume>2169</volume>
          {
          <fpage>2188</fpage>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Efthymios</surname>
            <given-names>Kouloumpis</given-names>
          </string-name>
          , Theresa Wilson, and
          <string-name>
            <given-names>Johanna</given-names>
            <surname>Moore</surname>
          </string-name>
          .
          <article-title>Twitter sentiment analysis: The good the bad and the OMG</article-title>
          !
          <source>In Proceedings of the Fifth International Conference on Weblogs and Social Media</source>
          ,
          <source>ICWSM '11</source>
          , pages
          <fpage>538</fpage>
          {
          <fpage>541</fpage>
          ,
          <string-name>
            <surname>Barcelona</surname>
          </string-name>
          , Catalonia, Spain,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <given-names>Saif</given-names>
            <surname>Mohammad</surname>
          </string-name>
          .
          <article-title>#emotional tweets</article-title>
          .
          <source>In *SEM 2012: The First Joint Conference on Lexical and Computational Semantics { Volume 1: Proceedings of the main conference and the shared task, and Volume 2: Proceedings of the Sixth International Workshop on Semantic Evaluation, SemEval '12</source>
          , pages
          <fpage>246</fpage>
          {
          <fpage>255</fpage>
          , Montreal, Canada,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Saif</surname>
            <given-names>Mohammad</given-names>
          </string-name>
          , Svetlana Kiritchenko, and
          <string-name>
            <given-names>Xiaodan</given-names>
            <surname>Zhu</surname>
          </string-name>
          .
          <article-title>NRC-Canada: Building the state-of-the-art in sentiment analysis of tweets</article-title>
          .
          <source>In Proceedings of the Seventh International Workshop on Semantic Evaluation Exercises</source>
          ,
          <source>SemEval '2013</source>
          , pages
          <fpage>321</fpage>
          {
          <fpage>327</fpage>
          ,
          <string-name>
            <surname>Atlanta</surname>
          </string-name>
          , Georgia, USA,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Preslav</surname>
            <given-names>Nakov</given-names>
          </string-name>
          , Sara Rosenthal, Zornitsa Kozareva, Veselin Stoyanov, Alan Ritter, and Theresa Wilson. SemEval
          <article-title>-2013 Task 2: Sentiment analysis in Twitter</article-title>
          .
          <source>In Proceedings of the Seventh International Workshop on Semantic Evaluation, SemEval '13</source>
          , pages
          <fpage>312</fpage>
          {
          <fpage>320</fpage>
          ,
          <string-name>
            <surname>Atlanta</surname>
          </string-name>
          , Georgia, USA,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Brendan O'Connor</surname>
            , Ramnath Balasubramanyan, Bryan Routledge, and
            <given-names>Noah</given-names>
          </string-name>
          <string-name>
            <surname>Smith</surname>
          </string-name>
          .
          <article-title>From tweets to polls: Linking text sentiment to public opinion time series</article-title>
          .
          <source>In Proceedings of the Fourth International Conference on Weblogs and Social Media</source>
          ,
          <source>ICWSM '10</source>
          , pages
          <fpage>122</fpage>
          {
          <fpage>129</fpage>
          , Washington, DC, USA,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <given-names>Alexander</given-names>
            <surname>Pak</surname>
          </string-name>
          and
          <string-name>
            <given-names>Patrick</given-names>
            <surname>Paroubek</surname>
          </string-name>
          .
          <article-title>Twitter based system: Using Twitter for disambiguating sentiment ambiguous adjectives</article-title>
          .
          <source>In Proceedings of the 5th International Workshop on Semantic Evaluation, SemEval '10</source>
          , pages
          <fpage>436</fpage>
          {
          <fpage>439</fpage>
          ,
          <string-name>
            <surname>Uppsala</surname>
          </string-name>
          , Sweden,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Bo</surname>
            <given-names>Pang</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Lillian</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>and Shivakumar</given-names>
            <surname>Vaithyanathan</surname>
          </string-name>
          .
          <article-title>Thumbs up?: Sentiment classi cation using machine learning techniques</article-title>
          .
          <source>In Proceedings of the Conference on Empirical Methods in Natural Language Processing -</source>
          Volume
          <volume>10</volume>
          , EMNLP '
          <volume>02</volume>
          , pages
          <fpage>79</fpage>
          {
          <fpage>86</fpage>
          ,
          <string-name>
            <surname>Philadephia</surname>
          </string-name>
          , Pennsylvania, USA,
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22. Francisco M. Rangel Pardo, Paolo Rosso, Moshe Koppel, Efstathios Stamatatos, and
          <string-name>
            <given-names>Giacomo</given-names>
            <surname>Inches</surname>
          </string-name>
          .
          <article-title>Overview of the author pro ling task at PAN 2013</article-title>
          . In Working Notes for CLEF 2013 Conference, Valencia, Spain,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Claudia</surname>
            <given-names>Peersman</given-names>
          </string-name>
          , Walter Daelemans, and Leona Van Vaerenbergh.
          <article-title>Predicting age and gender in online social networks</article-title>
          .
          <source>In Proceedings of the 3rd International Workshop on Search and Mining User-generated Contents, SMUC '11</source>
          , pages
          <fpage>37</fpage>
          {
          <fpage>44</fpage>
          , New York, NY, USA,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24. James W Pennebaker, Martha E Francis, and
          <string-name>
            <surname>Roger</surname>
          </string-name>
          J Booth.
          <article-title>Linguistic inquiry and word count: LIWC 2001</article-title>
          . Mahway: Lawrence Erlbaum Associates,
          <volume>71</volume>
          :
          <year>2001</year>
          ,
          <year>2001</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          25.
          <string-name>
            <given-names>F.</given-names>
            <surname>Rangel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          , and
          <string-name>
            <given-names>W.</given-names>
            <surname>Daelemans</surname>
          </string-name>
          .
          <article-title>Overview of the 3rd authorpro ling task at PAN 2015</article-title>
          . In Cappellato L.,
          <string-name>
            <surname>Ferro</surname>
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gareth</surname>
            <given-names>J.</given-names>
          </string-name>
          and San Juan E. (Eds). (Eds.)
          <article-title>CLEF 2015 Labs and Workshops</article-title>
          , Notebook Papers, Toulouse, France,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          26. Francisco Rangel, Paolo Rosso, Irina Chugur, Martin Potthast, Martin Trenkmann, Benno Stein, Ben Verhoeven, and
          <string-name>
            <given-names>Walter</given-names>
            <surname>Daelemans</surname>
          </string-name>
          .
          <article-title>Overview of the 2nd author pro ling task at PAN 2014</article-title>
          .
          <article-title>In CLEF 2014 Evaluation Labs</article-title>
          and Workshop { Working Notes Papers, She eld, UK,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          27.
          <string-name>
            <surname>Sara</surname>
            <given-names>Rosenthal</given-names>
          </string-name>
          , Alan Ritter, Preslav Nakov, and Veselin Stoyanov. SemEval
          <article-title>-2014 Task 9: Sentiment analysis in Twitter</article-title>
          .
          <source>In Proceedings of the 8th International Workshop on Semantic Evaluation, SemEval '14</source>
          , pages
          <fpage>73</fpage>
          {
          <fpage>80</fpage>
          , Dublin, Ireland,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          28.
          <string-name>
            <surname>Sara</surname>
            <given-names>Rosenthal</given-names>
          </string-name>
          , Alan Ritter, Veselin Stoyanov, Svetlana Kiritchenko, Saif Mohammad, and Preslav Nakov. SemEval-2015 task 10:
          <article-title>Sentiment analysis in Twitter</article-title>
          .
          <source>In Proceedings of the 9th International Workshop on Semantic Evaluation, SemEval '15</source>
          , pages
          <fpage>451</fpage>
          {
          <fpage>463</fpage>
          , Denver, CO, USA,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          29.
          <string-name>
            <given-names>Andrew</given-names>
            <surname>Schwartz</surname>
          </string-name>
          , Johannes Eichstaedt, Margaret Kern, Lukasz Dziurzynski, Stephanie Ramones, Megha Agrawal, Achal Shah, Michal Kosinski, David Stillwell,
          <string-name>
            <given-names>Martin</given-names>
            <surname>Seligman</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Lyle</given-names>
            <surname>Ungar</surname>
          </string-name>
          .
          <article-title>Personality, gender, and age in the language of social media: The open-vocabulary approach</article-title>
          .
          <source>In PLOS ONE</source>
          , volume
          <volume>8</volume>
          , page e73791.
          <source>Public Library of Science</source>
          , 09
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          30.
          <string-name>
            <surname>Janyce</surname>
            <given-names>Wiebe</given-names>
          </string-name>
          , Theresa Wilson, and
          <string-name>
            <given-names>Claire</given-names>
            <surname>Cardie</surname>
          </string-name>
          .
          <article-title>Annotating expressions of opinions and emotions in language</article-title>
          .
          <source>Language Resources and Evaluation</source>
          ,
          <volume>39</volume>
          (
          <issue>2- 3</issue>
          ):
          <volume>165</volume>
          {
          <fpage>210</fpage>
          ,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>