<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Spanish Twitter Messages Polarized through the Lens of an English System</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>Columbia University Columbia University Columbia University Columbia University New York</institution>
          ,
          <addr-line>NY USA New York, NY USA New York, NY USA New York, NY</addr-line>
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <fpage>81</fpage>
      <lpage>86</lpage>
      <abstract>
        <p>In this paper we describe the adaptation of a supervised classi cation system that was originally developed to detect sentiment on Twitter texts written in English. The Columbia University team adapted this system to participate in Task 1 of the 4th edition of the experimental evaluation workshop for sentiment analysis focused on the Spanish language (TASS 2015). The task consists of determining the global polarity of a group of messages written in Spanish using the social media platform Twitter.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Sentiment analysis is the eld concerned with
analyzing the sentimental content of text.
Most centrally, it involves the task of
deciding whether an utterance contains
subjectivity, as opposed to only objective statements,
and determining the polarity of such
subjective statements (e.g., whether the sentiment
is positive or negative). Automatic sentiment
analysis has important applications in
advertizing, social media, nance and other elds.
One variant that has become popular in
recent years is sentiment analysis in microblogs,
notably Twitter, which introduces di culties
common in that genre such as very short
utterances, non-standard language and frequent
out-of-vocabulary words.</p>
      <p>The vast majority of work on sentiment
analysis has been on English texts.
Since methods for determining sentiment often
rely on language-speci c resources such as
sentiment-tagged thesauri, they are often
difcult to adapt to languages beyond English,
as other language often have scarcer
computational resources.</p>
      <p>This paper describes the e orts of the
Columbia University team at Task 1 of TASS1
2015. TASS is an annual workshop focusing
on sentiment analysis in Spanish, especially
of short social media texts such as tweets.
Each year, TASS proposes a number of tasks
and collects the results of di erent
participating systems.</p>
      <p>
        In 2015, Task 1 is a combined
subjectivitypolarity task: for each tweet, the competing
system is expected to provide a label. There
are two variants - the ne-grained variant,
where there are six labels: fP+, P, Neu, N,
N+, NONEg, and the coarse variant, where
there are four labels: fP, Neu, N, NONEg.
TASS distributes a standard data set of over
68; 000 Spanish tweets for participants in this
task
        <xref ref-type="bibr" rid="ref17">(Villena-Roman et al., 2015)</xref>
        .
      </p>
      <p>Instead of creating a new Spanish-speci c
system, we have adapted our existing
English system to the Spanish language. We
show that with relatively small engineering
e orts and the proper resources, but without</p>
    </sec>
    <sec id="sec-2">
      <title>1Taller de Analisis de Sentimientos</title>
      <p>Publicado en http://ceur-ws.org/Vol-1397/. CEUR-WS.org es una publicación en serie con ISSN reconocido
any language-speci c feature engineering, our
system can be adapted to a new language and
achieve performance that is competitive with
other systems at TASS. As a side e ect, we
formalized the process of adapting our
system to any new language.
2</p>
      <sec id="sec-2-1">
        <title>Related Work</title>
        <p>
          Sentiment analysis in Twitter is a recent but
popular task. In English, the SemEval Task of
Sentiment Analysis in Twitter was the most
popular task in both 2013 and 2014
          <xref ref-type="bibr" rid="ref13 ref14">(Rosenthal et al., 2014)</xref>
          . In Spanish, TASS has
organized a Twitter sentiment analysis task every
year since 2012.
        </p>
        <p>
          Multiple papers focusing on this task have
been recently published. Most focus on
supervised classi cation, using lexical and
syntactic features
          <xref ref-type="bibr" rid="ref1 ref10 ref10 ref3 ref3 ref4 ref4 ref8">(Go, Bhayani, and Huang, 2009;
Pak and Paroubek, 2010; Bermingham and
Smeaton, 2010)</xref>
          . The latter, in particular,
compare polarity detection in twitter to the
same task in blogs, and nd that despite the
short and linguistically challenging nature of
tweets, it is easier to detect polarity in
Twitter than it is in blogs using lexical features,
presumably because of more sentimental
language in that medium.
        </p>
        <p>
          Other work focused on more specialized
features.
          <xref ref-type="bibr" rid="ref3">Barbosa and Feng (2010)</xref>
          use a
polarity dictionary that includes non-standard
(slang) vocabulary words as well as
Twitterspeci c social media features.
          <xref ref-type="bibr" rid="ref2">Agarwal et al.
(2011)</xref>
          use the Dictionary of A ect in
Language (DAL)
          <xref ref-type="bibr" rid="ref18">(Whissell, 1989)</xref>
          and social
media features such as slang and hashtags.
Rosenthal, McKeown, and Agarwal (2014) use
similar features, as well as features derived
from Wiktionary, WordNet and emoticon
dictionaries.
        </p>
        <p>
          In Spanish, most work on Twitter
sentiment analysis has been in the context of
TASS. Many of the top-performing systems
utilize a combination of lexical features, POS
and specialized lexicons: the Elhuyar system
relies on the Elhuyar Polar lexicon
          <xref ref-type="bibr" rid="ref11 ref13 ref16">(Roncal and Urizar, 2014)</xref>
          , while the LyS
system
          <xref ref-type="bibr" rid="ref11 ref13 ref16">(Vilares, Doval, and Gomez-Rodr guez,
2014)</xref>
          and the CITIUS-CILENIS system
          <xref ref-type="bibr" rid="ref12 ref5 ref7">(Gamallo and Garcia, 2013)</xref>
          each evaluate
several Spanish-language lexicons. Other systems
rely on distributional semantics
          <xref ref-type="bibr" rid="ref11 ref13 ref16">(MontejoRaez, Garcia-Cumbreras, and Diaz-Galiano,
2014)</xref>
          and on social media features
          <xref ref-type="bibr" rid="ref19 ref6">(Zafra et
al., 2014; Fernandez et al., 2013)</xref>
          .
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>Method</title>
        <p>
          The main e ort consisted of adapting an
English sentiment analysis system for Spanish
tweets, particularly for Task 1 of TASS 2015.
The English system has been successfully
applied to two editions of the SemEval
Task 9 (\Sentiment Analysis in Twitter")
- 2013 and 2014
          <xref ref-type="bibr" rid="ref13 ref14">(Rosenthal et al., 2014)</xref>
          .
The system consists of a Logistic Regression
classi er that utilizes a variety of lexical,
syntactic and specialized features (detailed
in Section 3.2). It has two modes that can
be run independently or in conjunction:
1. Subjectivity detection (distinguish
between subjective and objective tweets)
2. Polarity detection (classify subjective
tweets into positive, negative, or neutral).
        </p>
        <p>
          The system is described in detail in
          <xref ref-type="bibr" rid="ref12">Rosenthal and McKeown (2013)</xref>
          . For the
TASS task, four new modes were added:
1. Four-way classi cation, where the possible
classes are P, N, NEU, and NONE
2. Four-way composite classi cation, where
tweets are run through a two-step process:
a binary classi cation (subjective,
objective) followed by a three-way classi cation
(P,N,NEU) of subjective tweets. Objective
tweets in turn are given the label "NONE".
Consequently, this two-step classi cation
process yields to a four-way classi er. To
train the subjectivity classi er, we grouped
all labels other than \NONE" into one
subjective label.
3. Six-way classi cation, where the possible
classes are P, P+, N, N+, NEU, and NONE
4. Six-way composite classi cation (similar
to four-way composite, and including two
more labels: P+ and N+)
3.1
        </p>
        <sec id="sec-2-2-1">
          <title>Preprocessing of tweets</title>
          <p>
            Special tokens such as emoticons are replaced
by a related word (e.g. \smiley") and
supplemented with its a ect values as represented
in the DAL
            <xref ref-type="bibr" rid="ref18">(Whissell, 1989)</xref>
            . URLs and
Twitter handles are converted to xed tags that
are not analyzed further to determine
whether they are carriers of polarity. This process
is unchanged from the English system.
          </p>
          <p>We use the Stanford NLP library 2 for
tag2http://nlp.stanford.edu/software/index.shtml
ging and parsing the tweet. Using the parse
tree labels, we chunk the tweet into its
shallow syntactic constituents (e.g. grup.nom).
As in the English system, the chunker
outputs one of three labels per token to indicate
the position of the latter within a chunk: `B'
for beginning, `I' for in (or intermediate, a
continuation of the current chunk), and `O'
for out-of-vocabulary.
3.2</p>
        </sec>
        <sec id="sec-2-2-2">
          <title>Features</title>
          <p>The set of features used for Spanish is the
same as that of the English system; we did not
incorporate any Spanish-speci c features for
this task. The features currently utilized are
essentially those described in Rosenthal,
McKeown, and Agarwal (2014), and have
evolved over time from the original system
detailed in Agarwal, Biadsy, and Mckeown (2009).</p>
          <p>In addition to lexical features (n-grams
and POS), the system utilizes a variety of
specialized features for social media text:
emoticons; expanded web acronyms (LOL
! laugh out loud) and contractions (xq !
porque); punctuation and repeated
punctuation; lengthened words in the tweet (e.g.,
largooooooo); all-caps words; and slang. We also
use statistics of the DAL values for the words
in the tweet (e.g., the mean activation, the
max imagery, etc.).
3.3</p>
        </sec>
        <sec id="sec-2-2-3">
          <title>Adaptation to Spanish</title>
          <p>Adapting the English system to Spanish
included two parts. First, we had to nd
Spanish equivalents to the English lexical
resources (dictionaries, word lists etc.) that our
system relies on. Second, we had to nd
equivalent Spanish NLP tools (a tokenizer, POS
tagger and chunker).
3.3.1</p>
          <p>
            Lexical Resources
The major challenge we faced was the lack of
readily available resources in Spanish. In
some cases, Spanish resources could be found
and incorporated without a major e ort
for example, the Spanish version of the DAL
            <xref ref-type="bibr" rid="ref12 ref5 ref7">(Dell Amerlina R os and Gravano, 2013)</xref>
            was
simple to integrate. In other cases, we had to
put in more signi cant work - especially for
the social media resources (e.g. the lists of
contractions and emoticons). Table 1 details
the English lexical resources used by our
system and the Spanish equivalents, in addition
to the location in which we found them or the
method we used to create or adapt them.
          </p>
          <p>We integrated the Standard Spanish
dictionary distributed with Freeling3 as our
non-slang dictionary. For the DAL, we use
the Spanish version created by Dell
Amerlina R os and Gravano (2013). We leveraged
the Google Translate service to create a
Spanish version of our list of emoticons, and
manually created a list of Spanish contractions.</p>
          <p>The resulting Spanish resources are not
identical to the original English ones. For
example, the DAL scores for the word
\abandon" and its Spanish translation
\abandonar" are close but not exactly the same.
Furthermore, the number of entries in the
English DAL is more than three times that of
the Spanish one, which results in a signi cant
di erence in coverage. In the standard
dictionary, due to the highly in ected nature of
the Spanish language, the number of entries
more than quintuples when compared to the
English version. Table 2 shows the
percentage of the vocabulary (unique tokens) found
in the training corpus for each resource. The
standard dictionary has the highest coverage,
followed by the DAL.</p>
          <p>The English system utilizes a few
additional resources, namely Wiktionary, WordNet
and SentiWordNet. We have not yet
integrated a Spanish version of these into the
system, and consider that our rst priority in
future work. While Spanish equivalents of
Wiktionary and WordNet do exist (Wikcionario
and EuroWordNet), SentiWordNet does not
have non-English counterparts. Our planned
solution is to use MultiWordNet, a resource
in which the English WordNet is aligned with
other languages, to translate the English
Synsets included in SentiWordNet into Spanish.
# Found
10801</p>
          <p>Percentage
45.3 %
Resource
Standard
dictionary
DAL
NNP
Punctuation
and Numbers
Emoticons</p>
          <p>Tabla 2: Coverage of the training set
vocabulary by various resources</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3http://nlp.lsi.upc.edu/freeling</title>
      <p>Resource English
Dictionary of abandon
A ect in pleasantness mean: 1.0
Language (DAL) activation mean: 2.38
imagery mean: 2.4
. . . 8,742 entries
Contractions aren't ! are not
can't ! cannot
. . . 52 entries
Emoticons :) happy
:D laughter
:-( sad
. . . 99 entries
Standard 98,569 entries,
dictionary including proper
(i.e. not slang) nouns
Spanish
abandonar
1.2
2.8
2.0
. . . 2,669 entries
pal ! para el
palla ! para alla
. . . 21 entries
:) feliz
:D risa
:-( triste
. . . 99 entries
556,647 entries,
including in ected
forms</p>
      <p>Location or Method
http://habla.dc.uba.ar/gravano
/sdal.php?lang=esp
Manually created. Included variations
with and without accent marks,
apostrophes, and slang spelling.</p>
      <p>Translated using Google translate
Concatenated les contained in the
Freeling installation and removed
duplicates
Tabla 1: Parallel Lexical Resources
2000) for tagging. For chunking, we use the
Spanish version of the Stanford Parser, and
derive chunks from the lowermost syntactic
constituents (or the POS, if the token is not
a part of an immediate larger constituent).</p>
      <p>For example, the Spanish phrase \Buen
viernes" is chunked as follows:
Parse tree:
(ROOT (sentence (sn (grup.nom (s.a (grup.a
(aq0000 Buen))) (w viernes)))))
Chunked phrase:
Buen/aq0000/B-grup.a viernes/w/B-w
4</p>
      <sec id="sec-3-1">
        <title>Experiments and Results</title>
        <p>We submitted two experiments (one simple
and one composite; see Section 3) for each
combination of classi cation task (four-way,
six-way) and test corpus (full, 1k), for a total
of eight experiments. The results are shown in
Table 3. For each combination we show the
accuracy and the macro-averaged precision,
recall and F1 score.</p>
        <p>We trained ve models with the training
data provided by TASS:
1. Four-way (P, N, Neu, NONE)
2. Six-way (P+, P, N+, N, Neu, NONE)
3. Subjectivity model (subjective, objective)
4. Three-way polarity (P, N, Neu)
5. Five-way polarity (P+, P, N+, N, Neu)</p>
        <p>The last two were used in conjunction with
the subjectivity model to form the composite
classi er, as explained in Section 3.</p>
        <sec id="sec-3-1-1">
          <title>Discussion</title>
          <p>The task in which our system performs the
best is the three-label classi cation using
the joint four-way classi er described in
Section 3. Both joint models (four-way and
sixway) outperform their composite
counterparts on the full test corpus. However, there
is an improvement when using the composite
model on the balanced 1k corpus, for both
the three-label and ve-label classi cation.</p>
          <p>In terms of labels, our system consistently
has the most di culty classifying neutral
(Neu) tweets across all experiments. In
comparison, it did well in classifying strongly
positive (P+) and objective (NONE) tweets, as
well as positive (P) in the three-label
subtask. Negative (N, N+) tweets were in
between. Table 4 shows the performance of each
system (for each task) on individual labels.</p>
          <p>To assess the usefulness of our features
in discriminating among the di erent
classes, we looked at the odds ratios of the
features for each class. Table 5 shows a few of
the most discrimative features from each
category: n-grams, POS and social media (SM).</p>
          <p>We found that social media features
dominate across all classes, which is not a surprising
outcome given the popular use of such
features in Twitter communication. As shown
in Table 5, emoticons such as a smiley face
can be highly discriminative between
positive and negative tweets, with a signi cantly
stronger association with the former. Polar
N-grams such as \felices" (happy) also
constitute a relevant group and tend to be
discriminative for the polar classes N and P. In the</p>
          <p>POS group, interrogative pronouns (pt)
mar</p>
          <p>Tabla 3: Sentiment Analysis results at global level (all measures are macro-averaged)
Task
5 Labels</p>
          <p>TestSet</p>
          <p>Full
3 Labels
king words such as \que" (what) and \donde"
(where) are most important across all
categories, followed by various types of verbs
including semiauxiliary gerunds (vsg) and past
indicative auxiliary (vais).</p>
          <p>While it is di cult to compare our
system's Spanish results with the results on
English - the TASS dataset is quite di erent
from the SemEval dataset - it is evident that
the Spanish task is harder. This is not
surprising, since we have fewer resources, and
the ones which were adapted are in some
cases not as comprehensive. However, the fact
that we can get competitive results in
Spanish using a system that was originally
designed for English sentiment analysis shows
that relatively quick and painless adaptation
to other languages is possible.</p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>Conclusion</title>
        <p>We have adapted a sentiment analysis
system, the original target language of which
was English, to classifying the subjectivity
and polarity of tweets written in Spanish for
participation in Task 1 of TASS 2015. The
English system provided signi cant leverage,
allowing for direct reuse of most of its
components, from the processing pipeline down to
the features used by the classi er. The
experimental results are encouraging, showing our
system to be competitive with others
submitted to TASS despite being adapted into
Spanish from another language. From here on we
will pursue further enhancements.</p>
        <p>The main challenge we encountered was
the need to substitute several English
lexical resources that the system extensively
employs with analogous Spanish variants that
were not always easily attainable. In
future work, we will incorporate the nal
missing pieces - Spanish versions of Wiktionary,
WordNet and SentiWordNet - so that our
Spanish system uses equivalents of all
resources used by the English system.</p>
        <p>While adapting our system to Spanish,
we have compiled a list of necessary
resources and presented some automated methods
of quickly attaining such reasources in
other languages (e.g., using Google Translate
to quickly convert a list of emoticons). These
along with resources and tools that we expect
to be able to nd for most languages (e.g.,
a standard dictionary and a list of
contractions; a POS tagger and a constituent parser)
comprise the bulk of the list. Some resources,
such as the DAL, will potentially present a
bigger challenge in other languages, but can
possibly be automated through token
translation as well. In future work, we will
experiment with our system in additional languages
and further re ne our adaptation process.</p>
      </sec>
      <sec id="sec-3-3">
        <title>Acknowledgements</title>
        <p>This paper is based upon work supported by
the DARPA DEFT Program. The views
expressed are those of the authors and do not
re ect the o cial policy or position of the
Department of Defense or the U.S. Government.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <surname>Agarwal</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Biadsy</surname>
          </string-name>
          , and
          <string-name>
            <given-names>K. R.</given-names>
            <surname>McKeown</surname>
          </string-name>
          .
          <year>2009</year>
          .
          <article-title>Contextual phrase-level polarity analysis using lexical a ect scoring and syntactic n-grams</article-title>
          .
          <source>In EACL.</source>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>Agarwal</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Boyi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Vovsha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Rambow</surname>
          </string-name>
          , and
          <string-name>
            <given-names>R.</given-names>
            <surname>Passonneau</surname>
          </string-name>
          .
          <year>2011</year>
          .
          <article-title>Sentiment analysis of Twitter data</article-title>
          .
          <source>In Workshop on Languages in Social Media.</source>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <surname>Barbosa</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Feng</surname>
          </string-name>
          .
          <year>2010</year>
          .
          <article-title>Robust sentiment detection on Twitter from biased and noisy data</article-title>
          .
          <source>In COLING.</source>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <surname>Bermingham</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          and
          <string-name>
            <given-names>A. F.</given-names>
            <surname>Smeaton</surname>
          </string-name>
          .
          <year>2010</year>
          .
          <article-title>Classifying sentiment in microblogs: Is brevity an advantage?</article-title>
          <source>In International Conference on Information and Knowledge Management.</source>
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <surname>Dell Amerlina R os,</surname>
            <given-names>M.</given-names>
          </string-name>
          and
          <string-name>
            <given-names>A.</given-names>
            <surname>Gravano</surname>
          </string-name>
          .
          <year>2013</year>
          .
          <article-title>Spanish DAL: A Spanish dictionary of a ect in language.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <surname>Fernandez</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Gutierrez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Gomez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Martinez-Barco</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Montoyo</surname>
          </string-name>
          , and R. Mun~oz.
          <year>2013</year>
          .
          <article-title>Sentiment Analysis of Spanish Tweets Using a Ranking Algorithm and Skipgrams</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <surname>Gamallo</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          and
          <string-name>
            <surname>M. Garc</surname>
          </string-name>
          <article-title>a</article-title>
          .
          <year>2013</year>
          .
          <article-title>TASS: A Naive-Bayes strategy for sentiment analysis on Spanish tweets</article-title>
          .
          <source>In TASS</source>
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <surname>Go</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Bhayani</surname>
          </string-name>
          , and
          <string-name>
            <given-names>L.</given-names>
            <surname>Huang</surname>
          </string-name>
          .
          <year>2009</year>
          .
          <article-title>Twitter sentiment classi cation using distant supervision</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <surname>Montejo-Raez</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>M.A.</given-names>
            <surname>Garcia-Cumbreras</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.C.</given-names>
            <surname>Diaz-Galiano</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>Participacion de Sinai Word2vec en TASS 2014</article-title>
          . In TASS 2014.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <surname>Pak</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          and
          <string-name>
            <given-names>P.</given-names>
            <surname>Paroubek</surname>
          </string-name>
          .
          <year>2010</year>
          .
          <article-title>Twitter as a corpus for sentiment analysis and opinion mining</article-title>
          .
          <source>In Conference on International Lang. Resources and Evaluation.</source>
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <surname>Roncal</surname>
            ,
            <given-names>I. S. V.</given-names>
          </string-name>
          and
          <string-name>
            <given-names>X. S.</given-names>
            <surname>Urizar</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>Looking for features for supervised tweet polarity classi cation</article-title>
          .
          <source>In TASS</source>
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <surname>Rosenthal</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          and
          <string-name>
            <given-names>K.</given-names>
            <surname>McKeown</surname>
          </string-name>
          .
          <year>2013</year>
          .
          <string-name>
            <surname>Columbia</surname>
            <given-names>NLP</given-names>
          </string-name>
          :
          <article-title>Sentiment detection of subjective phrases in social media</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <surname>Rosenthal</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>K.</surname>
          </string-name>
          <article-title>McKeown, and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Agarwal</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <string-name>
            <surname>Columbia</surname>
            <given-names>NLP</given-names>
          </string-name>
          :
          <article-title>Sentiment detection of sentences and subjective phrases in social media</article-title>
          .
          <source>In SemEval.</source>
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <surname>Rosenthal</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ritter</surname>
          </string-name>
          , and
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>Semeval-2014 task 9: Sentiment analysis in Twitter.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <string-name>
            <surname>Toutanova</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          and
          <string-name>
            <given-names>C. D.</given-names>
            <surname>Manning</surname>
          </string-name>
          .
          <year>2000</year>
          .
          <article-title>Enriching the knowledge sources used in a maximum entropy part-of-speech tagger</article-title>
          .
          <source>In Joint SIGDAT Conference on Empirical Methods in NLP.</source>
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <surname>Vilares</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Doval</surname>
          </string-name>
          , and
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>GomezRodr guez</article-title>
          .
          <year>2014</year>
          . LyS at TASS 2014:
          <article-title>A prototype for extracting and analysing aspects from Spanish tweets</article-title>
          .
          <source>In TASS</source>
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <string-name>
            <surname>Villena-Roman</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Garc</surname>
          </string-name>
          a-Morera,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Garc</surname>
          </string-name>
          a-Cumbreras,
          <string-name>
            <surname>E.</surname>
          </string-name>
          <article-title>Mart nez-</article-title>
          <string-name>
            <surname>Camara</surname>
            ,
            <given-names>M. T.</given-names>
          </string-name>
          <string-name>
            <surname>Mart</surname>
            n-Valdivia, and
            <given-names>L. A.</given-names>
          </string-name>
          <string-name>
            <surname>UrenaLopez</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>Overview of TASS</article-title>
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <string-name>
            <surname>Whissell</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <year>1989</year>
          .
          <article-title>The dictionary of a ect in language</article-title>
          .
          <source>Emotion: Theory, research, and experience, 4.</source>
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <string-name>
            <given-names>Jimenez</given-names>
            <surname>Zafra</surname>
          </string-name>
          ,
          <string-name>
            <surname>S. M.</surname>
          </string-name>
          ,
          <string-name>
            <surname>E</surname>
          </string-name>
          . Mart nez Camara, M.T. Mart n Valdivia, and
          <string-name>
            <given-names>L.A.</given-names>
            <surname>Uren</surname>
          </string-name>
          ~a Lopez.
          <year>2014</year>
          .
          <article-title>Sinai-esma: An unsupervised approach for sentiment analysis in Twitter</article-title>
          .
          <source>In TASS</source>
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>