<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Sentiment Analysis for Twitter: TASS 2015</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Delegacion Tlalpan</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mexico D.F. osanchez</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>dmoctezuma@centrogeo.edu.mx</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Mario Gra Sabino Miranda-Jimenez Eric S. Tellez Elio-Atenogenes Villasen~or INFOTEC Ave. San Fernando 37</institution>
          ,
          <addr-line>Tlalpan, Toriello Guerra, 14050 Ciudad de Mexico, D.F. mario.gra ;sabino.miranda</addr-line>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2015</year>
      </pub-date>
      <volume>1397</volume>
      <fpage>65</fpage>
      <lpage>70</lpage>
      <abstract>
        <p>In this paper we present experiments for global polarity classi cation task of Spanish tweets for TASS 2015 challenge. In our methodology, tweets representation is focused on linguistic and polarity features such as lemmatized words, lter of content words, rules of negation, among others. In addition, di erent transformations are used (LDA, LSI, and TF-IDF) and combined with a SVM classi er. The results show that LSI and TF-IDF representations improve the performance of the SVM classi er applied.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        In last years the production of textual
documents in social media has increased
exponentially. This ever-growing amount of
available information promotes the research
and business activities around opinion
mining and sentiment analysis areas. In
social media, people share their opinions about
events, other people and organizations. This
is the main reason why text mining is
becoming an important research topic. Automatic
sentiment analysis in text is one of most
important task in text mining. The task of
sentiment classi cation determines if one
document has positive, negative or neutral
opinion or any level of each of them.
Determining whether a text document has a positive
or negative opinion is turning to an essential
tool for both public and private companies
        <xref ref-type="bibr" rid="ref7 ref9">(Peng, Zuo, y He, 2008)</xref>
        . This tool is
useful to know \What people think", which is
an important information in order to help to
any decision-making process (for any level of
government, marketing, etc.)
        <xref ref-type="bibr" rid="ref7 ref9">(Pang y Lee,
2008)</xref>
        . With this purpose, in this paper we
describe the methodology employed for the
workshop TASS 2015 (Taller de Analisis de
Sentimientos de la SEPLN). The TASS
workshop is an event of SEPLN conference, which
is a conference in Natural Language
Processing for Spanish language. The purpose of
TASS is to provide a discussion and a point of
sharing about latest research work in the eld
of sentiment analysis in social media (speci
cally Twitter in Spanish language). In TASS
workshop, several challenge tasks are
proposed, and furthermore a benchmark dataset
is proposed to compare the algorithms and
systems of participants (for more details see
        <xref ref-type="bibr" rid="ref12">(Villena-Roman et al., 2015)</xref>
        ).
      </p>
      <p>Several methodologies to classify tweets
from Task 1, Sentiment Analysis at global
level of TASS workshop 2015, are presented
in this work. This task is to perform an
automatic sentiment classi cation to determine
the global polarity (six polarity levels P, P+,
NEU, N, N+ and NONE) of each tweet in the
provided dataset. With this purpose, several
solutions have been proposed in this work.</p>
      <p>The paper is organized as follows, a brief
overview of related works is shown in Section
2, the proposed methodology is describe in
Section 3. Section 4 shows the
experimental results and analysis, and nally, Section
5 concludes.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Related work</title>
      <p>Nowadays, several methods have been
proposed in the community of opinion mining
and sentiment analysis. Most of these works
employ twitter as a principal input of data
and they aimed to classify entire documents
as overall positive or negative polarity levels
(sentiment) or rating scores (i.e. 1 to 5 stars).</p>
      <p>
        Such is a case of work presented in
        <xref ref-type="bibr" rid="ref11 ref2">(da
Silva, Hruschka, y Jr., 2014)</xref>
        , which proposes
an approach to classify sentiment of tweets by
using classi er ensembles and lexicons; where
tweets are classi ed as positive or negative.
As a result, this work concludes that
classier ensembles formed by several and diverse
components are promising for tweet
sentiment classi cation. Moreover, several
stateof-the-art techniques were compared in four
databases. The best accuracy result reported
was around 75%.
      </p>
      <p>
        In
        <xref ref-type="bibr" rid="ref2 ref5">(Llu s F. Hurtado, 2014)</xref>
        is described
the participation of ELiRF research group in
TASS 2014 workshop (winners of TASS
workshop 2014). Here, the winner approaches
used for four tasks are detailed. The
proposed methodology uses SVM (Support
Vector Machines) with 1-vs-all approach.
Moreover, Freeling
        <xref ref-type="bibr" rid="ref6">(Padro y Stanilovsky, 2012)</xref>
        was used as lemmatizer and Tweetmotif1 to
tokenizer to Spanish language. The accuracy
results of classi cation for task 1 are 64:32%
(six labels) and 70:89% (four labels). F1
(FMeasure) is 70:48% in task 2 and 90% in task
3.
      </p>
      <p>
        Another method to sentiment extraction
and classi cation on unstructured text is
proposed in
        <xref ref-type="bibr" rid="ref11">(Shahbaz, Guergachi, y ur Rehman,
2014)</xref>
        . Here, ve labels were used to
sentiment classi cation: Strongly Positive,
Positive, Neutral, Negative and Strongly
Negative. The solution proposed combines
techniques of Natural language processing at
sentence level and algorithms of opinion mining.
The accuracy results were 61% for ve levels
and 75% by reducing to three levels (Positive,
negative and neutral).
      </p>
      <p>
        In
        <xref ref-type="bibr" rid="ref1">(Antunes et al., 2011)</xref>
        an ensemble
based on SVM and AIS (Arti cial Immune
Systems) is proposed. Here, the main idea
is that SVM can be enhanced with AIS
approaches which can capture dynamic
models. Experiments were carried out with the
Reuters-21578 benchmark dataset. The
reported results show a 95:52% of F1.
      </p>
      <p>
        An approach of multi-label sentiment
classi cation is proposed in
        <xref ref-type="bibr" rid="ref4">(Liu y Chen,
2015)</xref>
        . This approach has three main
components: text segmentation, feature
extraction and multi-label classi cation. The
features used included raw segmented words
and sentiment features based on three
sentiment dictionaries: DUTSD, NTUSD and
HD. Moreover, here, a detailed study of
several multi-label classi cation methods is
conducted, in total 11 state-of-the-art
methods have been considered: BR, CC, CLR,
HOMER, RAkEL, ECC, MLkNN, and
RFPCT, BRkNN, BRkNN-a and BRkNN-b.
These methods were compared in two
microblog datasets and the reported results of
all methods are around of 0:50 of F1.
      </p>
      <p>In summary, most of works analyzed
classify the documents mainly in three
polarities: positive, neutral and negative.
Moreover, most of works use social media (mainly
Twitter) as analyzed documents. In this
work, several methods to classify sentiment
in tweets are described. These methods were
implemented, according with TASS
workshop speci cations, with the purpose of
classify tweets in six polarity levels: P+, P,
Neutral, N+, N and None. The proposed method
are based on several standard techniques as
LDA (Latent Dirichlet Allocation), LSI
(Latent Semantic Indexing), TF-IDF matrix in
tom):
combination with the well-known SVM
classi er.
3</p>
    </sec>
    <sec id="sec-3">
      <title>Proposed solution</title>
      <p>In this section the proposed solution is
detailed. First, a preprocessing step was
carried out, later a Pseudo-phonetic
transformation was done and nally the generation of
Q-gram expansion was employed.
3.1</p>
      <sec id="sec-3-1">
        <title>Preprocessing step</title>
        <p>
          Preprocessing focuses on the task of
nding a good representation for tweets. Since
tweets are full of slang and misspellings, we
normalize the text using procedures such as
error correction, usage of special tags, part
of speech (POS) tagging, and negation
processing. Error correction consists on
reducing words/tokens with invalid duplicate
vowels and consonants to valid/standard
Spanish words (ruidoooo ! ruido; jajajaaa ! ja;
jijijji ! ja). Error correction uses an
approach based on a Spanish dictionary,
statistical model for common double letters, and
heuristic rules for common interjections. In
the case of the usage of special tags, twitter's
users (i.e., @user) and urls are removed
using regular expressions; in addition, we
classify 512 popular emoticons into four classes
(P, N, NEU, NONE), which are replaced by a
polarity tag in the text, e.g., positive
emoticons such as :), :D are replaced by POS,
and negative emoticons such as :(, :S are
replaced by NEG. In the POS-tagging step, all
words are tagged and lemmatized using the
Freeling tool for Spanish language
          <xref ref-type="bibr" rid="ref6">(Padro y
Stanilovsky, 2012)</xref>
          , stop words are removed,
and only content words (nouns, verbs,
adjetives, adverbs), interjections, hashtags, and
polarity tags are used for data representation.
In negation step, Spanish negation markers
are attached to the nearest content word, e.g.,
`no seguir' is replaced by `no seguir', `no es
bueno' is replaced by `no bueno', `sin comida'
is replaced by `no comida'; we use a set of
heuristic rules for negations. Finally, all
diacritic and punctuation symbols are also
removed.
3.2
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>Psudo-phonetic transformation</title>
        <p>With the purpose of reducing typos and
slangs we applied a semi-phonetic
transformation. First, we applied the following
transformations (with precedence from top to
botcsjxc ! x</p>
        <p>qu ! k
guejge ! je
guijgi ! ji
shjch ! x
ll ! y
z ! s
h !
c[ajoju] ! k
c[eji] ! s
w ! u
v ! b
!
!</p>
        <p>
          In our transformation notation, square
brackets do not consume symbols and ;
means for any valid symbols. The idea is
not to produce a pure phonetic
transformation as in Soundex
          <xref ref-type="bibr" rid="ref3">(Donald, 1999)</xref>
          like
algorithms, but try to reduce the number of
possible errors in the text. Notice that the
last two transformation rules are partially
covered by the statistical modeling used for
correcting words (explained in
preprocessing step). Nonetheless, this pseudo-phonetic
transformation does not follow the statistical
rules of the previous preprocessing step.
3.3
        </p>
      </sec>
      <sec id="sec-3-3">
        <title>Q-gram expansion</title>
        <p>Along with the placing bag of words
representation (of the normalized text) we added
the 4 and 5 gram of characters of the
normalized text. Blank spaces were normalized
and taken into account to the q-gram
expansion; so, some q-grams will be over more than
one word. In addition of these previous steps,
several transformations (LSI, LDA and
TFIDF matrix) were conducted to generate
several data models for testing phase.
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Results and analysis</title>
      <p>The classi er submitted to the competition
was selected using the following procedure.</p>
      <p>The 7; 218 tweets with 6 polarity levels were
split in two sets. Firstly, the tweets
provided were shu ed and then the rst set,
hereafter the training set, was created with
the rst 6; 496 tweets (approximately 90% of
dataset), and, the second set, hereafter the
validation set, was composed by the rest 722
tweets (approximately 10% of dataset). The
training set was used to t a Support Vector
Machine (SVM) using a linear kernel2 with
C = 1, weights inversely proportional to the
class frequencies, and using one vs rest
multiclass strategy. The validation set was used to
select the best classi er using as performance
the score F1.</p>
      <p>
        The rst step was to model the data
using di erent transformations, namely Latent
Dirichlet Allocation (LDA) using an online
learning proposed by
        <xref ref-type="bibr" rid="ref10 ref4">(Ho man, Bach, y Blei,
2010)</xref>
        , Latent Semantic Indexing (LSI), and
TF-IDF.3 Figure 1 presents the score F1, in
the validation set, of a SVM using either LSI
or LDA with normalized text, di erent
levels of Q-gram (4-gram and 5-gram), and the
number of topics is varied from 10 to 500 as
well. It is observed that LSI outperformed
LDA in all the con gurations tested.
Comparing the performance between normalized
text, 4-gram, and 5-gram, it is observed an
equivalent performance. Given that the
implemented LSI depends on the order of the
documents more experiments are needed to
know whether any particular con guration is
statistically better than other. Even though
the best con guration is LSI with 400 topics
and 5-gram, this system is not competitive
enough compared with the performance
presented by the best algorithm in TASS 2014.
      </p>
      <p>Table 1 complements the information
presented on Figure 1.</p>
      <p>The table presents the score F1 per
polarity and the average (Macro-F1) for di erent
con gurations. The table is divided in ve
blocks, the rst and second correspond to a
SVM with LSI (400 topics) and TF-IDF,
respectively. It is observed that TF-IDF
outperformed LSI; within LSI and TF-IDF it
can be seen that 5-gram and 4-gram got the
best performance in LSI and TF-IDF,
respectively.</p>
      <p>The third row block presents the
performance when the features are a direct addition
of LSI and TF-IDF; here it is observed that
the best performance is with 4-gram
furthermore it had the best overall performance in
N+. The forth row block complements the
previous results by presenting the best
performance of LSI and TF-IDF, that is, LSI
with 5-gram and TF-IDF with 4-gram. It
is observed that this con guration has the
best overall performance in P+, N, None and
average (Macro-F1). Finally, the last row
block gives an indicated of whether the
phonetic transformation is making any
improvement. The conclusion is that the phonetic
transformation is making a di erence;
however, more experiments are needed in order
to know whether this di erence is statistically
signi cant.</p>
      <p>Based on the score F1 presented on Table
1 the classi er submitted to the competition
is a SVM with a direct addition of LSI using
400 topics and 4-gram and LDA with 5-gram.</p>
      <p>This classi er is identi ed as
INGEOTECM14 in the competition. The SVM, LSI and
LDA were trained with the 7218 tweets and
then this instance was used to predict the
6 polarity levels of the competition tweets.</p>
      <p>This procedure was replicated for the 4
polarity levels competition.</p>
      <p>Table 4 presents the accuracy, average
recall, precision, and F1 of INGEOTEC-M1
run using the validation set created, a
10fold crossvalidation on the 7218 tweets and
the 1k tweets evaluated by the system's
competition. This performance was on the 5
polarity levels challenge. It is observed from the
table that the 10-fold crossvalidation gives a
much better estimation of the performance</p>
      <p>4We also submitted another classi er identi ed as
INGEOTEC-E1; however, the algorithm presented a
bug that could not be nd out on time for the
competition.
P P+ N N+ Neutral None</p>
      <p>SVM + LSI</p>
      <p>Text 0:238 0:549 0:403 0:348 0:025 0:492
4-gram 0:246 0:543 0:404 0:333 0:048 0:533
5-gram 0:246 0:552 0:462 0:356 0:000 0:575</p>
      <p>SVM + TF-IDF</p>
      <p>Text 0:271 0:574 0:414 0:407 0:103 0:511
4-gram 0:290 0:577 0:477 0:393 0:130 0:589
5-gram 0:302 0:577 0:476 0:379 0:040 0:586</p>
      <p>SVM + fLSI + TF-IDFg
4-gram 0:297 0:578 0:471 0:421 0:142 0:578
5-gram 0:307 0:567 0:474 0:391 0:040 0:579</p>
      <p>SVM + fLSI with 4-gram + TF-IDF with 5-gramg
4-5-gram 0:282 0:596 0:481 0:407 0:144 0:595</p>
      <p>SVM + fLSI + TF-IDF without phonetic transformationg
4-5-gram 0:324 0:577 0:459 0:395 0:150 0:593
0:343
0:351
0:365
of the classi er when tested on 1k tweets of
the competition (90% of training and 10% of
validation).</p>
      <p>In summary, in this work the best result
reached was a 0.404 of F1. This result was
achieved with a combination of LSI with
4gram + TF-IDF with 5-gram, using a SVM
classi er (one-vs-one approach).</p>
      <p>Val.
10-fold
Comp.</p>
      <p>Acc.
0:471
0:443
0:431</p>
      <p>Recall
0:428
0:397
0:411</p>
      <p>Precision
0:421
0:395
0:398</p>
      <p>F1
0:417
0:393
0:404
In this contribution, we presented the
approach used to tackle the polarity classi
cation task of Spanish tweets of TASS 2015.</p>
      <p>From the results, it is observed that a
combination of di erent data models, in this case
LSI and TF-IDF, improves the performance
of a SVM classi er. It also noted that the
phonetic transformation makes an
improvement; however, more experiments are needed
to know whether this improvement is
statistically signi cant. As a result, we obtained a
0.404 of F1 (macro-F1) in sentiment classi
cation task at ve levels, with the proposed
solution. This proposed solution uses a
combination of LSI with 4-gram + TF-IDF with
5-gram, and a SVM classi er (one-vs-one
approach).</p>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgements</title>
      <p>This research is partially supported by the
Catedras CONACyT project. Furthermore,
the authors would like to thank CONACyT
for supporting this work through the project
247356 (PN2014).
in neural information processing systems,
paginas 856{864.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <surname>Antunes</surname>
          </string-name>
          , Mario, Catarina Silva, Bernardete Ribeiro, y Manuel Correia.
          <year>2011</year>
          .
          <article-title>A hybrid ais-svm ensemble approach for text classi cation</article-title>
          .
          <source>En Adaptive and Natural Computing Algorithms, volumen 6594 de Lecture Notes in Computer Science</source>
          . Springer Berlin Heidelberg, paginas
          <volume>342</volume>
          {
          <fpage>352</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>da Silva</surname>
            , Nadia
            <given-names>F.F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Eduardo</surname>
            <given-names>R</given-names>
          </string-name>
          . Hruschka,
          <string-name>
            <given-names>y Estevam R. Hruschka</given-names>
            <surname>Jr</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>Tweet sentiment analysis with classi er ensembles</article-title>
          .
          <source>Decision Support Systems</source>
          ,
          <volume>66</volume>
          (
          <issue>0</issue>
          ):
          <volume>170</volume>
          {
          <fpage>179</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <surname>Donald</surname>
            ,
            <given-names>E</given-names>
          </string-name>
          <string-name>
            <surname>Knuth</surname>
          </string-name>
          .
          <year>1999</year>
          .
          <article-title>The art of computer programming</article-title>
          .
          <source>Sorting and searching</source>
          ,
          <volume>3</volume>
          :
          <fpage>426</fpage>
          {
          <fpage>458</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <surname>Ho</surname>
            <given-names>man</given-names>
          </string-name>
          , Matthew, Francis R Bach, y David M Blei.
          <year>2010</year>
          .
          <article-title>Online learning for latent dirichlet allocation</article-title>
          .
          <source>En advances Liu</source>
          , Shuhua Monica y
          <string-name>
            <surname>Jiun-Hung Chen</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>A multi-label classi cation based approach for sentiment classi cation</article-title>
          .
          <source>Expert Systems with Applications</source>
          ,
          <volume>42</volume>
          (
          <issue>3</issue>
          ):
          <volume>1083</volume>
          {
          <fpage>1093</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <surname>Llu</surname>
            s
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Hurtado</surname>
            ,
            <given-names>Ferran</given-names>
          </string-name>
          <string-name>
            <surname>Pla</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>Elirfupv en TASS</article-title>
          <year>2014</year>
          : Analisis de sentimientos, deteccion de topicos y analisis de sentimientos de aspectos en twitter.
          <source>Proc. of the TASS workshop at SEPLN</source>
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <surname>Padro</surname>
          </string-name>
          , Llu s y Evgeny Stanilovsky.
          <year>2012</year>
          .
          <article-title>Freeling 3.0: Towards wider multilinguality</article-title>
          .
          <source>En Proceedings of the Language Resources and Evaluation Conference (LREC</source>
          <year>2012</year>
          ), Istanbul, Turkey, May. ELRA.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <surname>Pang</surname>
          </string-name>
          , Bo y Lillian Lee.
          <year>2008</year>
          .
          <article-title>Opinion mining and sentiment analysis</article-title>
          .
          <source>Foundations and Trends in Information Retrieval</source>
          ,
          <volume>2</volume>
          (
          <issue>1</issue>
          - 2):1{
          <fpage>135</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <surname>Pedregosa</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Varoquaux</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gramfort</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Michel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Thirion</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Grisel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Blondel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Prettenhofer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Weiss</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Dubourg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Vanderplas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Passos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Cournapeau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Brucher</surname>
          </string-name>
          , M. Perrot,
          <string-name>
            <given-names>y E.</given-names>
            <surname>Duchesnay</surname>
          </string-name>
          .
          <year>2011</year>
          .
          <article-title>Scikit-learn: Machine learning in Python</article-title>
          .
          <source>Journal of Machine Learning Research</source>
          ,
          <volume>12</volume>
          :
          <fpage>2825</fpage>
          {
          <fpage>2830</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <surname>Peng</surname>
          </string-name>
          , Tao, Wanli Zuo, y Fengling He.
          <year>2008</year>
          .
          <article-title>Svm based adaptive learning method for text classi cation from positive and unlabeled documents</article-title>
          .
          <source>Knowledge and Information Systems</source>
          ,
          <volume>16</volume>
          (
          <issue>3</issue>
          ):
          <volume>281</volume>
          {
          <fpage>301</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <surname>Rehurek</surname>
          </string-name>
          , Radim y Petr Sojka.
          <year>2010</year>
          .
          <article-title>Software Framework for Topic Modelling with Large Corpora</article-title>
          .
          <source>En Proceedings of the LREC 2010 Workshop on New Challenges for NLP Frameworks</source>
          , paginas
          <volume>45</volume>
          {
          <fpage>50</fpage>
          ,
          <string-name>
            <surname>Valletta</surname>
          </string-name>
          , Malta, Mayo. ELRA.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <surname>Shahbaz</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          . Guergachi, y R.
          <source>T. ur Rehman</source>
          .
          <year>2014</year>
          .
          <article-title>Sentiment miner: A prototype for sentiment analysis of unstructured data and text</article-title>
          .
          <source>En Electrical and Computer Engineering (CCECE)</source>
          ,
          <source>2014 IEEE 27th Canadian Conference on, paginas</source>
          <volume>1</volume>
          {7, May.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <surname>Villena-Roman</surname>
            , Julio,
            <given-names>Janine</given-names>
          </string-name>
          <string-name>
            <surname>Garc</surname>
          </string-name>
          a-Morera,
          <article-title>Miguel A. Garc a-Cumbreras, Eugenio Mart nez-</article-title>
          <string-name>
            <surname>Camara</surname>
          </string-name>
          , M. Teresa Mart nValdivia, y L.
          <article-title>Alfonso Uren~a-</article-title>
          <string-name>
            <surname>Lopez</surname>
          </string-name>
          .
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <surname>Overview of TASS</surname>
          </string-name>
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>