<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>ELiRF-UPV at IberEval 2017: Classi cation Of Spanish Election Tweets (COSET)</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Jose-Angel Gonzalez</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ferran Pla</string-name>
          <email>fpla@dsic.upv.es</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Llu s-F. Hurtado</string-name>
          <email>lhurtado@dsic.upv.es</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Departament de Sistemes Informatics i Computacio Universitat Politecnica de Valencia</institution>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2017</year>
      </pub-date>
      <fpage>55</fpage>
      <lpage>60</lpage>
      <abstract>
        <p>This paper describes the participation of ELiRF-UPV team at Classi cation Of Spanish Election Tweets (COSET) task. We tested several approaches based on di erent classi ers and features representations. Our main approach is based on neural networks, concretely, Multilayer Perceptrons (MLP) with bag-of-words representation of the tweets. Our system achieved the best score on the test set of the COSET task with 64.82 of macro-F1.</p>
      </abstract>
      <kwd-group>
        <kwd>Neural Networks</kwd>
        <kwd>MLP</kwd>
        <kwd>bag-of-words</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Text Categorization (TC) is a well-known and widely applied machine-learning
technique for classifying textual documents into categories [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. The goal of TC
is the classi cation of documents into a xed number of prede ned categories.
      </p>
      <p>
        Recently, there have been several works that adapt traditional TC methods
for classifying Twitter posts into a prede ned set of generic classes [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
      </p>
      <p>
        The aim of COSET shared task at Ibereval 2017 [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] is to classify a corpus of
political tweets in ve categories: political issues related to electoral
confrontation, sectoral policies, personal questions about the candidates, issues of the
electoral campaign and other issues. This data has been collected during the
Spanish electoral campaign of 2015.
      </p>
      <p>The present paper describes the participation of the ELiRF-UPV team at
COSET shared task, the way in which the task has been addressed and the
results obtained using di erent approaches.
The corpus is composed by tweets belonging to ve categories: political issues,
related to the most abstract electoral confrontation; policy issues, about sectorial
policies; personal issues, on the life and activities of the candidates; campaign
issues, related with the evolution of the campaign; and other issues.</p>
      <p>These tweets are written in Spanish and were collected during the general
Spanish elections of 2015. The organization provided this corpus divided in three
parts. The training set (2242 tweets), the development set (250 tweets) and the
test set (624 tweets).</p>
      <p>Note that the topic distribution of the corpus is not uniform (see Table 1).
The three majority classes (political issues, policy issues and personal issues)
represent more than the 80% of the samples. Moreover, both training and
development sets have similar distributions of tweets per category.
3</p>
    </sec>
    <sec id="sec-2">
      <title>System Description</title>
      <p>In this section we describe the main characteristics of the system developed to the
COSET task competition. This description includes the preprocessing used, the
feature selection process and the di erent models that were taken into account
during the tuning phase.
3.1</p>
      <sec id="sec-2-1">
        <title>Preprocessing</title>
        <p>In the preprocessing conducted all the tweets were converted to lowercase and the
accents were removed. In all cases, Url's, user's mentions, numbers, exclamations
and interrogations are replaced by a speci c label. On the contrary, hashtags were
not replaced because we observed that hashtags were relevant in the classi cation
process.
3.2</p>
      </sec>
      <sec id="sec-2-2">
        <title>Models and Feature Selection</title>
        <p>
          Di erent classi ers and tweets representations were considered. Three models
were tested: Support Vector Machines (SVM) with linear kernel, Multilayer
Perceptron (MLP), and deep learning models (CNN+LSTM), that we have recently
used in Sentiment Analysis tasks [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ].
        </p>
        <p>
          We selected the most appropriate tweets representations for each considered
model. This way, we tested SVM and MLP models with bag-of-ngrams (n 1) of
words and characters and TF-IDF weighting; and CNN+LSTM with sequences
of Spanish Wikipedia embeddings [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ] [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ] and sequences of one-hot
vectors.
        </p>
        <p>The best results were obtained using Multilayer Perceptron (MLP) models
with bag-of-words (BOW) as representation of tweets. Support Vector Machines
also presented good results, in particular by using BOW representations.
Contrary to what we expected, deep learning models obtained the worst results.</p>
        <p>Speci cally, our best model was a Multilayer Perceptron of 3 hidden layers
with 128 neurons and ReLU as activation function in all layers (except Softmax
in the output layer).</p>
        <p>
          The optimization method was Adagrad [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ], which has provided the best
results in similar TC tasks, and categorical cross-entropy as a loss function. In
addition, a scaling of the loss function is used to deal with the problem of
unbalanced classes. This scaling method has the following form: floss(x) log( nnrc )
where nr is the number of samples of the majority class and nc is the number
of samples of the class of sample x.
4
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Results</title>
      <p>In view of the results obtained during the tuning phase, we decided to send four
systems with the same base architecture (Multilayer Perceptron with
bag-ofwords representation), but with some minor changes:
{ System-1. A MLP trained using all data (training and development sets)
with the parameters adjusted on the development set.
{ System-2. A MLP trained using only the training set.
{ System-3. A MLP trained using with all data, but without scaling the loss
function.
{ System-4. A majority voting schema of the three previous systems.</p>
      <p>The o cial results achieved by our four systems are shown in Table 3. We
have also included, in parenthesis, the position of each system in the ranking of
the COSET competition.</p>
      <p>The best results of the competition were obtained by our System-1, which
achieved 64.82 of macro-F1 on the test set. The scaling of the loss function
adjusted on the development set also behaved well when the system is trained
with all samples.</p>
      <p>System-2 obtained the worst results of our submissions (5th position over 39
submissions). Note that, this was the system learned with less data. System-3 was
learned with the same samples of System-1 but without scaling the loss function.
From the results achieved by System-3, we want to highlight the importance of
dealing with the problem of unbalanced classes.</p>
      <p>The majority voting proposed (System-4) could not outperform the results of
the best model (System-1). Perhaps, a more sophisticated combination method
could obtain better results.</p>
      <p>Finally, our team has participated in the second evaluation phase proposed
by the COSET organization. In this second phase, a new corpus was
automatically labeled with the agreement of four of the ve participant runs (80% of
agreement). A 65.91% of the considered tweets achieved the agreement
criterion. Therefore, the nal corpus size was 10,417,058 tweets. Table 4 shows the
results of the second evaluation phase, sorted by macro-F1 measure. We achieved
the best two results in terms of macro-F1 in this second phase. In this case, the
best system, ELiRF-UPV.1, was trained with the whole COSET corpus and the
second system, ELiRF-UPV.2, was learned using only the training set of the
COSET corpus.
5</p>
    </sec>
    <sec id="sec-4">
      <title>Conclusions and Future Work</title>
      <p>We have presented the participation of ELiRF-UPV team at Classi cation Of
Spanish Election Tweets (COSET) task. We tested several approaches. The best
results were obtained by Multilayer Perceptrons models with bag-of-words
representation. Our system achieved the best score on the competition with 64.82
of macro-F1.</p>
      <p>In this task, the scaling of the loss function seems to be a key factor to
improve the results, as it can be observed in the nal ranking, where the system
that uses this scaling performs better than the systems that do not use it.</p>
      <p>
        As future work, we plan to study the generation of synthetic samples to tackle
with the unbalance problem. We want to test data-augmentation techniques and
generative models as SMOTE [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], GAN [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], or VAE [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgements</title>
      <p>This work has been partially supported by the Spanish MINECO and FEDER
founds under project ASLP-MULAN: Audio, Speech and Language Processing
for Multimedia Analytics, TIN2014-54288-C4-3-R.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Bowyer</surname>
            ,
            <given-names>K.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chawla</surname>
            ,
            <given-names>N.V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hall</surname>
            ,
            <given-names>L.O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kegelmeyer</surname>
            ,
            <given-names>W.P.:</given-names>
          </string-name>
          <article-title>SMOTE: synthetic minority over-sampling technique</article-title>
          .
          <source>CoRR abs/1106</source>
          .
          <year>1813</year>
          (
          <year>2011</year>
          ), http://arxiv. org/abs/1106.1813
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Duchi</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hazan</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Singer</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>Adaptive subgradient methods for online learning and stochastic optimization</article-title>
          .
          <source>J. Mach. Learn. Res</source>
          .
          <volume>12</volume>
          ,
          <issue>2121</issue>
          {2159 (Jul
          <year>2011</year>
          ), http: //dl.acm.org/citation.cfm?id=
          <volume>1953048</volume>
          .
          <fpage>2021068</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Fiaidhi</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mohammed</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Islam</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fong</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kim</surname>
          </string-name>
          , T.h.:
          <article-title>Developing a hierarchical multi-label classi er for twitter trending topics</article-title>
          .
          <source>International Journal</source>
          of U-&amp;
          <string-name>
            <surname>E-Service</surname>
          </string-name>
          ,
          <source>Science &amp; Technology</source>
          <volume>6</volume>
          (
          <issue>3</issue>
          ) (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Gimenez</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Baviera</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Llorca</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gamir</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Calvo</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rangel</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>Overview of the 1st Classi cation of Spanish Election Tweets Task at IberEval 2017</article-title>
          .
          <source>In: Proceedings of the Second Workshop on Evaluation of Human Language Technologies for Iberian Languages (IberEval</source>
          <year>2017</year>
          ).
          <source>CEUR Workshop Proceedings. CEUR-WS.org, Murcia</source>
          , Spain (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Gonzalez</surname>
            ,
            <given-names>J.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pla</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hurtado</surname>
            ,
            <given-names>L.F.</given-names>
          </string-name>
          :
          <article-title>ELiRF-UPV at SemEval-2017 Task 4: Sentiment Analysis using Deep Learning</article-title>
          .
          <source>In: Proceedings of the 11th International Workshop on Semantic Evaluation</source>
          . pp.
          <volume>722</volume>
          {
          <fpage>726</fpage>
          . SemEval '17,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computational Linguistics, Vancouver, Canada (
          <year>August 2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Goodfellow</surname>
            ,
            <given-names>I.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pouget-Abadie</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mirza</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xu</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Warde-Farley</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ozair</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Courville</surname>
            ,
            <given-names>A.C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bengio</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>Generative adversarial networks</article-title>
          .
          <source>CoRR abs/1406</source>
          .2661 (
          <year>2014</year>
          ), http://arxiv.org/abs/1406.2661
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Kingma</surname>
            ,
            <given-names>D.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Welling</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Auto-encoding variational bayes</article-title>
          .
          <source>CoRR abs/1312</source>
          .6114 (
          <year>2013</year>
          ), http://arxiv.org/abs/1312.6114
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Palsetia</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Narayanan</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Patwary</surname>
            ,
            <given-names>M.M.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Agrawal</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Choudhary</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Twitter trending topic classi cation</article-title>
          .
          <source>In: Proceedings of the 2011 IEEE 11th International Conference on Data Mining Workshops</source>
          . pp.
          <volume>251</volume>
          {
          <fpage>258</fpage>
          . ICDMW '11, IEEE Computer Society, Washington, DC, USA (
          <year>2011</year>
          ), http://dx.doi.org/10. 1109/ICDMW.
          <year>2011</year>
          .171
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Corrado</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dean</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>E cient estimation of word representations in vector space</article-title>
          .
          <source>CoRR abs/1301</source>
          .3781 (
          <year>2013</year>
          ), http://arxiv.org/abs/ 1301.3781
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sutskever</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Corrado</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dean</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          :
          <article-title>Distributed representations of words and phrases and their compositionality</article-title>
          .
          <source>CoRR abs/1310</source>
          .4546 (
          <year>2013</year>
          ), http://arxiv.org/abs/1310.4546
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Rehurek</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sojka</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Software Framework for Topic Modelling with Large Corpora</article-title>
          .
          <source>In: Proceedings of the LREC 2010 Workshop on New Challenges for NLP Frameworks</source>
          . pp.
          <volume>45</volume>
          {
          <fpage>50</fpage>
          . ELRA, Valletta, Malta (May
          <year>2010</year>
          ), http://is.muni.cz/ publication/884893/en
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Sebastiani</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <article-title>: Machine learning in automated text categorization</article-title>
          .
          <source>ACM Comput. Surv</source>
          .
          <volume>34</volume>
          (
          <issue>1</issue>
          ),
          <volume>1</volume>
          {
          <fpage>47</fpage>
          (Mar
          <year>2002</year>
          ), http://doi.acm.
          <source>org/10</source>
          .1145/505282.505283
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13. Wikipedia:
          <article-title>Wikipedia spanish dumps (</article-title>
          <year>2017</year>
          ), https://dumps.wikimedia.org/ eswiki/, [Online; accessed 18-May-2017]
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>