<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>On the performance of B4MSA on SENTIPOLC'16</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Col. Tecnopolo Pocitos II</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Me´xico dmoctezuma@centrogeo.edu.mx</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Edmund K Burke</institution>
          ,
          <addr-line>Graham Kendall, et al. Search methodologies. Springer</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Eric S. Tellez Mario Graff Sabino Miranda-Jime ́nez CONACyT-INFOTEC Circuito Tecnopolo Sur No 112, Fracc. Tecnopolo Pocitos II</institution>
          ,
          <addr-line>Ags, 20313, Me ́xico</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>This document describes the participation of the INGEOTEC team in SENTIPOLC 2016 contest. In this participation two approaches are presented, B4MSA and B4MSA + EvoDAG, tested in Task 1: Subjectivity classification and Task 2: Polarity classification. In case of polarity classification, one constrained and unconstrained runs were conducted. In subjectivity classification only a constrained run was done. In our methodology we explored a set of techniques as lemmatization, stemming, entity removal, character-based q-grams, word-based n-grams, among others, to prepare different text representations, in this case, applied to the Italian language. The results show the official competition measures and other well-known performance measures such as macro and micro F1 scores.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Italiano. Questo documento descrive
la partecipazione del team INGEOTEC
alla competizione SENTIPOLC 2016. In
questo contributo sono presentati due
approcci, B4MSA e B4MSA + EvoDAG,
applicati al Task 1: Subjectivity
classification e Task 2: Polarity classification. Nel
caso della classificazione della polarit,
sono stati sottomessi un run constrained
ed un run unconstrained. Per la
classificazione della soggettivita, stato
sottomesso solo un run constrained. La
nostra metodologia esplora un insieme di
tecniche come lemmatizzazione, stemming,
rimozione di entit, q-grammi di caratteri,
n-grammi di parole, ed altri, al fine di
ottenere diverse rappresentazioni del testo.
In questo caso essa applicata alla
lingua italiana. I risultati qui presentati sono
due: le metriche della competizione
ufficiale ed altre misure note della
performance, come macro F1 e micro F1.</p>
    </sec>
    <sec id="sec-2">
      <title>1 Introduction</title>
      <p>Nowadays, the sentiment analysis task has become
a problem of interest for governments, companies,
and institutions due to the possibility of sensing
massively the mood of the people using social
networks in order to take advantage in
decisionmaking process. This new way to know what are
people thinking about something imposes
challenges to the natural language processing and
machine learning areas, the first of all, is that
people using social networks are kindly ignoring
formal writing. For example, a typical Twitter user
do not follow formal writing rules and introduces
new lexical variations indiscriminately, the use of
emoticons and the mix of languages is also the
common lingo. These characteristics produce high
dimensional representations, where the curse of
dimension makes hard to learn from examples.</p>
      <p>There exists a number of strategies to cope with
the sentiment analysis on Twitter messages, some
of them are based on the fact that the core problem
is fixed: we are looking for evidence of some
sentiment in the text. Under this scheme a number of
dictionaries have been described by psychologists,
other resources like SentiWordNet have been
created adapting well known linguistic resources and
machine learning. There is a lot of work around
this approach; however, all these knowledge is
language dependent and must exists a deep
understanding of the language being analyzed. Our
approach is mostly independent of this kind of
external resources while focus on tackling the
misspellings and other common errors in the text.</p>
      <p>In this manuscript we detail our approach to
sentiment analysis from a language agnostic
perspective, e.g., no one in our team knows Italian
language. We neither use external knowledge nor
specialized parsers. Our aim is to create a solid
baseline from a multilingual perspective, that can
be used as a real baseline for challenges like
SENTIPOLC’16 and as a basic initial approximation
for sentiment analysis systems.</p>
      <p>The rest of the paper is organized in the
following sections. Section 2 describes our approach.
Section 3 describes our experimental results, and
finally Section 4 concludes.
2</p>
    </sec>
    <sec id="sec-3">
      <title>Our participation</title>
      <p>This participation is based on two approaches.
First, B4MSA method, a simple approach which
starts by applying text-transformations to the
tweets, then transformed tweets are represented in
a vector space model, and finally, a Support Vector
Machine (with linear kernel) is used as the
classifier. Second, B4MSA + EvoDAG, a combination
of this simple approach with a Genetic
programming scheme.
2.1</p>
      <sec id="sec-3-1">
        <title>Text modeling with B4MSA</title>
        <p>B4MSA is a system for multilingual polarity
classification that can serve as a baseline as well as a
framework to build sophisticated sentiment
analysis systems due to its simplicity. The source code
of B4MSA can be downloaded freely1.</p>
        <p>We used our previous work, B4MSA, to tackle
the SENTIPOLC challenge. Our approach learns
based on training examples, avoiding any digested
knowledge as dictionaries or ontologies. This
scheme allows us to address the problem without
caring about the particular language being tackled.</p>
        <p>
          The dataset is converted to a vector space using
a standard procedure: the text is normalized,
tokenized and weighted. The weighting process is
fixed to be performed by TFIDF
          <xref ref-type="bibr" rid="ref1">(Baeza-Yates and
Ribeiro-Neto, 2011)</xref>
          . After that process, a linear
SVM (Support Vector Machines) is trained using
10-fold cross-validation (Burges, 1998). At the
end, this classifier is applied to the test set to
obtain the final prediction.
        </p>
        <p>1https://github.com/INGEOTEC/b4msa</p>
        <p>At a glance, our goal is to find the best
performing normalization and tokenization pipelines. We
state the modeling as a combinatorial
optimization problem; then, given a performance measure,
we try to find the best performing configuration
among a large parameter space.</p>
        <p>
          The list of transformations and tokenizers are
listed below. All the text transformations
considered are either simple to implement, or there is
an open-source library (e.g.
          <xref ref-type="bibr" rid="ref9">(Bird et al., 2009;
Rˇehu˚rˇek and Sojka, 2010)</xref>
          ) that implement it.
2.2
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>Set of Features</title>
        <p>In order to find the best performing
configuration, we used two sort of features that we consider
them as parameters: cross-language and
languagedependent features.</p>
        <p>
          Cross-language Features could be applied in
most similar languages and similar surface
features. Removing or keeping punctuation
(question marks, periods, etc.) and diacritics from the
original source; applying or not applying the
processes of case sensitivity (text into lowercase) and
symbol reduction (repeated symbols into one
occurrence of the symbol). Word-based n-grams
(nwords) Feature are word sequences of words
according to the window size defined. To compute
the N-words, the text is tokenized and combined
the tokens. For example, 1-words (unigrams) are
each word alone, and its 2-words (bigrams) set are
the sequences of two words, and so on
          <xref ref-type="bibr" rid="ref6">(Jurafsky and Martin, 2009)</xref>
          . Character-based q-grams
(q-grams) are sequences of characters. For
example, 1-grams are the symbols alone, 3-grams are
sequences of three symbols, generally, given text
of size m characters, we obtain a set with at most
m q + 1 elements
          <xref ref-type="bibr" rid="ref8">(Navarro and Raffinot, 2002)</xref>
          .
Finally, Emoticon (emo) feature consists in
keeping, removing, or grouping the emotions that
appear in the text; popular emoticons were hand
classified (positive, negative or neutral), included text
emoticons and the set of unicode emoticons
          <xref ref-type="bibr" rid="ref13">(Unicode, 2016)</xref>
          .
        </p>
      </sec>
      <sec id="sec-3-3">
        <title>Language Dependent Features. We considered</title>
        <p>
          three language dependent features: stopwords,
stemming, and negation. These processes are
applied or not applied to the text. Stopwords
and stemming processes use data and the
Snowball Stemmer for Italian, respectively, from NLTK
Python package (Bird et al., 2009). Negation
feature markers could change the polarity of the
message. We used a set of language dependent rules
for common negation structures to attached the
negation clue to the nearest word, similar to the
approach used in
          <xref ref-type="bibr" rid="ref11">(Sidorov et al., 2013)</xref>
          .
2.3
        </p>
      </sec>
      <sec id="sec-3-4">
        <title>Model Selection</title>
        <p>The model selection, sometimes called
hyperparameter optimization, is the key of our
approach. The default search space of B4MSA
contains more than 331 thousand configurations when
limited to multilingual and language independent
parameters; while the search space reaches close
to 4 million configurations when we add our three
language-dependent parameters. Depending on
the size of the training set, each configuration
needs several minutes on a commodity server to
be evaluated; thus, an exhaustive exploration of
the parameter space can be quite expensive that
makes the approach useless.</p>
        <p>
          To reduce the selection time, we perform a
stochastic search with two algorithms, random
search and hill climbing. Firstly, we apply
random search
          <xref ref-type="bibr" rid="ref4">(Bergstra and Bengio, 2012)</xref>
          that
consists on randomly sampling the parameter space
and select the best configuration among the
sample. The second algorithm consists on a hill
climbing
          <xref ref-type="bibr" rid="ref3">(Burke et al., 2005; Battiti et al., 2008)</xref>
          implemented with memory to avoid testing a
configuration twice. The main idea behind hill
climbing is to take a pivoting configuration (in our
case we start using the best one found by random
search), explore the configuration’s neighborhood,
and greedily moving to the best neighbor. The
process is repeated until no improvement is possible.
The configuration neighborhood is defined as the
set of configurations such that these differ in just
one parameter’s value.
        </p>
        <p>Finally, the performance of the final
configuration is obtained applying the above procedure and
cross-validation over the training data.
2.4</p>
      </sec>
      <sec id="sec-3-5">
        <title>B4MSA + EvoDAG</title>
        <p>
          In the polarity task besides submitting B4MSA
which is a constrained approach, we decided to
generate an unconstrained submission by
performing the following approach. The idea is to
provide an additional dataset that it is automatically
label with positive and negative polarity using the
Distant Supervision approach
          <xref ref-type="bibr" rid="ref12 ref7">(Snow et al., 2005;
Morgan et al., 2004)</xref>
          .
        </p>
        <p>We start collecting tweets (using Twitter
stream) written in Italian. In total, we collect
more than 10; 000; 000 tweets. From these tweets,
we kept only those that were consistent with the
emoticon’s polarity used, e.g., the tweet only
contains consistently emoticons with positive polarity.
Then, the polarity of the whole tweet was set to the
polarity of the emoticons, and we only used
positive and negative polarities. Furthermore, we
decided to balance the set, and then we remove a lot
of positive tweets. At the end, this external dataset
contains 4; 550; 000 tweets, half of them are
positive and the another half are negative.</p>
        <p>Once this external dataset was created, we
decided to split it in batches of 50; 000 tweets half
of them positive and the other half negative. This
decision was taken in order to optimize the time
needed to train a SVM and also around this
number the Macro F1 metric is closed to its maximum
value. That is, this number of tweets gives a good
trade-off between time needed and classifier
performance. In total there are 91 batches.</p>
        <p>For each batch, we train a SVM at the end of
this process we have 91 predictions (it is use the
decision function). Besides these 91 predictions, it
is also predicted (using as well the decision
function) each tweet with B4MSA. That is, at the end
of this process we have 94 values for each tweet.
That is, we have a matrix with 7; 410 rows and
94 columns for the training set and of 3; 000 rows
and 94 columns for the test set. Moreover, for
matrix of the training set, we also know the class for
each row. It is important to note that all the
values of these matrix are predicted, for example, in
B4MSA case, we used a 10-fold cross-validation
in the training set in order to have predicted values.</p>
        <p>
          Clearly, at this point, the problem is how to
make a final prediction; however, we had built
a classification problem using the decision
functions and the classes provided by the competition.
Thus, it is straight forward to tackle this
classification problem using EvoDAG (Evolving Directed
Acyclic Graph)2
          <xref ref-type="bibr" rid="ref5">(Graff et al., 2017)</xref>
          which is a
Genetic Programming classifier that uses
semantic crossover operators based on orthogonal
projections in the phenotype space. In a nutshell,
EvoDAG was used to ensemble the outputs of the
91 SVM trained with the dataset automatically
labeled and B4MSA’s decision functions.
        </p>
        <p>2https://github.com/mgraffg/EvoDAG</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Results and Discussion</title>
      <p>
        This Section presents the results of the
INGEOTEC team. In this participation we did two
runs, a constrained and an unconstrained run with
B4MSA system, and only a constrained run with
B4MSA + EvoDAG. The constrained run was
conducted only with the dataset provided by
SENTIPOLC’16 competition. For more technical
details from the database and the competition in
general see
        <xref ref-type="bibr" rid="ref2">(Barbieri et al., 2016)</xref>
        .
      </p>
      <p>
        The unconstrained run was developed with an
additional dataset of 4; 550; 000 of tweets labeled
with Distant Supervision approach. The Distant
Supervision is an extension of the paradigm used
in
        <xref ref-type="bibr" rid="ref12">(Snow et al., 2005)</xref>
        and nearest to the use of
weakly labeled data in
        <xref ref-type="bibr" rid="ref7">(Morgan et al., 2004)</xref>
        . In
this case, we consider the emoticons as key for
automatic labeling. Hence, a tweet with a high
level of positive emoticons is labeled as positive
class and a tweet with a clear presence of negative
emoticons is labeled as negative class. This give
us a bigger amount of samples for the dataset for
training.
      </p>
      <p>
        For the constrained run we participate in two
task: subjectivity and polarity classification. In
the unconstrained run we only participate in
polarity classification task. Table 1 shows the results of
subjectivity classification Task (B4MSA method),
here, Prec0 is the P recision0 value, Rec0 is the
Recall0 value, FSc0 is F Score0 value and
Prec1, Rec1 and FSc1 the same for F Score1
values and FScavg is the average value from all
F-Scores. The explanation of evaluation measures
can be seen in
        <xref ref-type="bibr" rid="ref2">(Barbieri et al., 2016)</xref>
        .
      </p>
      <p>Table 2, shows the results on the polarity
classification task. In this task our B4MSA method
achieves an average F-Score of 0:6054 and our
combination of B4MSA + EvoDAG reaches an
0:6075 of average F-Score. These results place us
on position 18 (unconstrained run) and 19
(constrained run) of a total of 26 entries.</p>
      <p>It is important to mention that the difference
between our two approaches is very small; however,
B4MSA + EvoDAG is computationally more
expensive, so we expected to have a considerable
improvement in performance. It is evident that
these results should be investigated further, and,
our first impression are that our Distant
supervision approach should be finely tune, that is, it is
needed to verify the polarity of the emoticons and
the complexity of the tweets.</p>
      <p>
        Finally, Table 3 presents the measures
employed by our internal measurement, that is Macro
F1 and Micro F1 (for more details see
        <xref ref-type="bibr" rid="ref10">(Sebastiani,
2002)</xref>
        ). These values are from polarity
unconstrained run (B4MSA + EvoDAG), polarity
constrained run (B4MSA), subjectivity constrained
run (B4MSA) and irony classification (B4MSA).
We do not participate in irony classification task
but we want to show the obtained result from our
B4MSA approach on this task.
4
      </p>
    </sec>
    <sec id="sec-5">
      <title>Conclusions</title>
      <p>In this work we describe the INGEOTEC team
participation in SENTIPOLC’16 contest. Two
approaches were used, first, B4MSA method which
combine several text transformations to the tweets.
Secondly, B4MSA + EvoDAG, which combine the
B4MSA method with a genetic programming
approach. In subjectivity classification task, the
obtained results place us in seventh of a total of 21
places. In polarity classification task, our results
place us 18 and 19 places of a total of 26. Since
our approach is simple and easy to implement, we
take these results important considering that we do
not use affective lexicons or another complex
linguistic resource. Moreover, our B4MSA approach
was tested internally in irony classification task
with a result of 0:4687 of macro f1, and 0:8825
of micro f1.
0.56
FScorepos
0.6414</p>
      <sec id="sec-5-1">
        <title>FScoreneg Combined FScore</title>
      </sec>
      <sec id="sec-5-2">
        <title>Constrained run (B4MSA)</title>
        <p>0.5694 0.6054</p>
      </sec>
      <sec id="sec-5-3">
        <title>Unconstrained run (B4MSA + EvoDAG)</title>
        <p>0.5944 0.6205 0.6075
Steven Bird, Ewan Klein, and Edward Loper.
2009. Natural Language Processing with Python.</p>
        <p>O’Reilly Media.</p>
        <p>Christopher J.C. Burges. 1998. A tutorial on support
vector machines for pattern recognition. Data
Mining and Knowledge Discovery, 2(2):121–167.
2005.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <given-names>Ricardo A.</given-names>
            <surname>Baeza-Yates</surname>
          </string-name>
          and
          <article-title>Berthier A</article-title>
          .
          <string-name>
            <surname>Ribeiro-Neto</surname>
          </string-name>
          .
          <year>2011</year>
          .
          <article-title>Modern Information Retrieval. AddisonWesley, 2nd edition</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <given-names>Francesco</given-names>
            <surname>Barbieri</surname>
          </string-name>
          , Valerio Basile, Danilo Croce, Malvina Nissim, Nicole Novielli, and
          <string-name>
            <given-names>Viviana</given-names>
            <surname>Patti</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Overview of the EVALITA 2016 SENTiment POLarity Classification Task</article-title>
          . In Pierpaolo Basile, Anna Corazza, Franco Cutugno, Simonetta Montemagni, Malvina Nissim, Viviana Patti, Giovanni Semeraro, and Rachele Sprugnoli, editors,
          <source>Proceedings of Third Italian Conference on Computational Linguistics</source>
          (CLiC-it
          <year>2016</year>
          ) &amp;
          <article-title>Fifth Evaluation Campaign of Natural Language Processing and Speech Tools for Italian</article-title>
          .
          <source>Final Workshop (EVALITA</source>
          <year>2016</year>
          ).
          <article-title>Associazione Italiana di Linguistica Computazionale (AILC).</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <given-names>Roberto</given-names>
            <surname>Battiti</surname>
          </string-name>
          , Mauro Brunato, and
          <string-name>
            <given-names>Franco</given-names>
            <surname>Mascia</surname>
          </string-name>
          .
          <year>2008</year>
          .
          <article-title>Reactive search and intelligent optimization</article-title>
          , volume
          <volume>45</volume>
          . Springer Science &amp; Business Media.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <given-names>James</given-names>
            <surname>Bergstra</surname>
          </string-name>
          and
          <string-name>
            <given-names>Yoshua</given-names>
            <surname>Bengio</surname>
          </string-name>
          .
          <year>2012</year>
          .
          <article-title>Random search for hyper-parameter optimization</article-title>
          .
          <source>Journal of Machine Learning Research</source>
          ,
          <volume>13</volume>
          (Feb):
          <fpage>281</fpage>
          -
          <lpage>305</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <given-names>Mario</given-names>
            <surname>Graff</surname>
          </string-name>
          , Eric S. Tellez, Hugo Jair Escalante, and
          <string-name>
            <surname>Sabino</surname>
          </string-name>
          Miranda-Jimnez.
          <year>2017</year>
          .
          <article-title>Semantic Genetic Programming for Sentiment Analysis</article-title>
          . In Oliver Schtze, Leonardo Trujillo, Pierrick Legrand, and Yazmin Maldonado, editors,
          <source>NEO</source>
          <year>2015</year>
          ,
          <article-title>number</article-title>
          663
          <source>in Studies in Computational Intelligence</source>
          , pages
          <fpage>43</fpage>
          -
          <lpage>65</lpage>
          . Springer International Publishing.
          <source>DOI: 10.1007/978-3-319-44003-3 2.</source>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <given-names>Daniel</given-names>
            <surname>Jurafsky and James H. Martin</surname>
          </string-name>
          .
          <year>2009</year>
          .
          <article-title>Speech and Language Processing (2Nd Edition)</article-title>
          .
          <source>PrenticeHall</source>
          , Inc., Upper Saddle River, NJ, USA.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <given-names>Alexander A.</given-names>
            <surname>Morgan</surname>
          </string-name>
          , Lynette Hirschman, Marc Colosimo, Alexander S. Yeh, and
          <string-name>
            <surname>Jeff</surname>
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Colombe</surname>
          </string-name>
          .
          <year>2004</year>
          .
          <article-title>Gene name identification and normalization using a model organism database</article-title>
          .
          <source>Journal of Biomedical Informatics</source>
          ,
          <volume>37</volume>
          (
          <issue>6</issue>
          ):
          <fpage>396</fpage>
          -
          <lpage>410</lpage>
          . Named Entity Recognition in Biomedicine.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <given-names>G.</given-names>
            <surname>Navarro</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Raffinot</surname>
          </string-name>
          .
          <year>2002</year>
          .
          <article-title>Flexible Pattern Matching in Strings - Practical on-line search algorithms for texts and biological sequences</article-title>
          . Cambridge University Press.
          <source>ISBN 0-521-81307-7</source>
          . 280 pages.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <article-title>Radim Rˇ ehu˚rˇek</article-title>
          and
          <string-name>
            <given-names>Petr</given-names>
            <surname>Sojka</surname>
          </string-name>
          .
          <year>2010</year>
          .
          <article-title>Software Framework for Topic Modelling with Large Corpora</article-title>
          .
          <source>In Proceedings of the LREC 2010 Workshop on New Challenges for NLP Frameworks</source>
          , pages
          <fpage>45</fpage>
          -
          <lpage>50</lpage>
          , Valletta, Malta, May. ELRA. http://is. muni.cz/publication/884893/en.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <given-names>Fabrizio</given-names>
            <surname>Sebastiani</surname>
          </string-name>
          .
          <year>2002</year>
          .
          <article-title>Machine learning in automated text categorization</article-title>
          .
          <source>ACM Comput. Surv.</source>
          ,
          <volume>34</volume>
          (
          <issue>1</issue>
          ):
          <fpage>1</fpage>
          -
          <lpage>47</lpage>
          ,
          <year>March</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <given-names>Grigori</given-names>
            <surname>Sidorov</surname>
          </string-name>
          , Sabino Miranda-Jime´nez, Francisco Viveros-Jime´nez, Alexander Gelbukh, Noe´ CastroSa´nchez, Francisco Vela´squez, Ismael D´
          <article-title>ıaz-</article-title>
          <string-name>
            <surname>Rangel</surname>
          </string-name>
          ,
          <article-title>Sergio Sua´rez-</article-title>
          <string-name>
            <surname>Guerra</surname>
          </string-name>
          ,
          <source>Alejandro Trevin˜o, and Juan Gordon</source>
          .
          <year>2013</year>
          .
          <article-title>Empirical study of machine learning based approach for opinion mining in tweets</article-title>
          .
          <source>In Proceedings of the 11th Mexican International Conference on Advances in Artificial Intelligence - Volume Part I, MICAI'12</source>
          , pages
          <fpage>1</fpage>
          -
          <lpage>14</lpage>
          , Berlin, Heidelberg. Springer-Verlag.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <surname>Rion</surname>
            <given-names>Snow</given-names>
          </string-name>
          , Daniel Jurafsky, and
          <string-name>
            <given-names>Andrew Y.</given-names>
            <surname>Ng</surname>
          </string-name>
          .
          <year>2005</year>
          .
          <article-title>Learning syntactic patterns for automatic hypernym discovery</article-title>
          . In L. K. Saul,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Weiss</surname>
          </string-name>
          , and L. Bottou, editors,
          <source>Advances in Neural Information Processing Systems</source>
          <volume>17</volume>
          , pages
          <fpage>1297</fpage>
          -
          <lpage>1304</lpage>
          . MIT Press.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <surname>Unicode</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Unicode emoji chart</article-title>
          . http://unicode.org/emoji/charts/ full-emoji-list.
          <source>html. Accessed 20-May2016.</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>