<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Series</journal-title>
      </journal-title-group>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Neural Networks for Sentiment Analysis in Czech</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ladislav Lenc</string-name>
          <email>llenc@kiv.zcu.cz</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Tomáš Hercig</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science and Engineering, Faculty of Applied Sciences, University of West Bohemia</institution>
          ,
          <addr-line>Univerzitní 8, 306 14 Plzenˇ</addr-line>
          ,
          <country country="CZ">Czech Republic</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>NTIS-New Technologies for the Information Society, Faculty of Applied Sciences, University of West Bohemia</institution>
          ,
          <addr-line>Technická 8, 306 14 Plzenˇ</addr-line>
          ,
          <country>Czech</country>
          <addr-line>Republic nlp.kiv.zcu.cz</addr-line>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2016</year>
      </pub-date>
      <volume>1649</volume>
      <fpage>48</fpage>
      <lpage>55</lpage>
      <abstract>
        <p>This paper presents the first attempt at using neural networks for sentiment analysis in Czech. The neural networks have shown very good results on sentiment analysis in English, thus we adapt them to the Czech environment. We first perform experiments on two English corpora to allow comparability with the existing state-ofthe-art methods for sentiment analysis in English. Then we explore the effectiveness of using neural networks on four Czech corpora. We show that the networks achieve promising results however there is still much room for improvement especially on the Czech corpora.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1 Introduction</title>
      <p>
        The current approaches to sentiment analysis in English
explore various neural network architectures (e.g. [
        <xref ref-type="bibr" rid="ref1 ref2 ref3">1, 2,
3</xref>
        ]). We try to replicate the results shown in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] and adapt
the proposed architecture to the sentiment analysis task in
Czech – a highly inflectional Slavic language. To the best
of our knowledge, neural networks have not been used for
the sentiment analysis task in Czech.
      </p>
      <p>The goal of aspect-based sentiment analysis (ABSA) is
to identify the aspects of a given target entity and estimate
the sentiment polarity for each mentioned aspect, while the
general goal of sentiment analysis is to detect the polarity
of a text. In this work we will focus on polarity detection
on various levels (texts, sentences, and aspects).</p>
      <p>
        In recent years the aspect-based sentiment analysis has
undergone rapid development mainly because of
competitive tasks such as SemEval 2014 - 2016 [
        <xref ref-type="bibr" rid="ref4 ref5 ref6">4, 5, 6</xref>
        ].
      </p>
      <p>Aspect-based sentiment analysis firstly identifies the
aspects of target entity and then assigns a polarity to each
aspect. There are several ways to define aspects and
polarities.</p>
      <p>The definition of the ABSA task from SemEval 2014
distinguishes two types of aspect-based sentiment: aspect
terms and aspect categories. The whole task is divided into
four subtasks.</p>
      <p>• Aspect Term Extraction (TE) – identify aspect
terms.</p>
      <p>Our server checked on us maybe twice during the
entire meal.
→ {server, meal}
• oAfsepaeccht TasepremctPteorlmar.ity (TP) – determine the polarity
Our server checked on us maybe twice during the
entire meal.</p>
      <p>→ {server: negative, meal: neutral}
• fiAnsepde)ctasCpaetcetgcoarteygEoxritersa.ction (CE) – identify
(predeOur server checked on us maybe twice during the
entire meal.</p>
      <p>→ {service}
• Aspect Category Polarity (CP) – determine the
polarity of each (pre-identified) aspect category.</p>
      <p>Our server checked on us maybe twice during the
entire meal.</p>
      <p>→ {service: negative}</p>
      <p>The later SemEval’s ABSA tasks (2015 and 2016)
further distinguish between more detailed aspect categories
and associate aspect terms (targets) with aspect categories.</p>
      <p>
        The current ABSA task - SemEval 2016 [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] has three
subtasks: Sentence-level (SB1), Text-level (SB2) and
Outof-domain ABSA (SB3). The subtasks are further divided
into three slots. The following example is from the training
data (including the typographical error).
      </p>
      <p>• 1fi)neAd)spaescptecCt actaetgeogroyryD–eteencttiitoynan–d
iadtetnritbifuyte(p(Ere#dAe)pair.</p>
      <p>The pizza is yummy and I like the atmoshpere.
→ {FOOD#QUALITY, AMBIENCE#GENERAL}
• 2) Opinion Target Expression (OTE) – extract the
OTE referring to the reviewed entity (aspect
category).</p>
      <p>The pizza is yummy and I like the atmoshpere.
→ {pizza, atmoshpere}
• 3n)egSaetinvtei,maenndt nPeoultararli)tyto–eaacsshigindenptoilfiaerdityE#(pAo,siOtiTveE,
tuple.</p>
      <p>The pizza is yummy and I like the atmoshpere.
→ {FOOD#QUALITY - pizza: positive,</p>
      <p>AMBIENCE#GENERAL - atmoshpere: positive}</p>
      <p>In this work we will focus on the sentiment polarity
task on aspect-level and document-level1 for Czech and
English. In terms of the SemEval 2014 task it is the
Aspect Term Polarity and Aspect Category Polarity (TP and
CP) subtasks. In terms of the SemEval 2016 task it is the
Sentence-level Sentiment Polarity subtask.</p>
      <p>Our main goal is to measure the difference between the
previous results and the new results achieved by neural
network architectures.
2
2.1</p>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <sec id="sec-2-1">
        <title>Sentiment Analysis in Czech</title>
        <p>
          Initial research on Czech sentiment analysis has been done
in [
          <xref ref-type="bibr" rid="ref10 ref7 ref8 ref9">7, 8, 9, 10</xref>
          ]. However they used only small news
datasets and because of the small data size no strong
conclusions can be drawn.
        </p>
        <p>
          The first extensive evaluation of Czech sentiment
analysis was done by Habernal et al. in [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ]. Three different
classifiers, namely Naive Bayes, Support Vector Machines
and Maximum Entropy classifiers were tested on
largescale labeled corpora (Facebook posts, movie reviews, and
product reviews). In [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] they further experimented with
feature selection methods.
        </p>
        <p>
          Habernal and Brychcin [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ] used semantic spaces (see
[
          <xref ref-type="bibr" rid="ref14">14</xref>
          ]) created from unlabeled data as an additional source
of information to improve results. Brychcin and Habernal
[
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] explored the benefits of the global target context and
outperformed the previous unsupervised approach.
        </p>
        <p>
          The first attempt at aspect-based sentiment analysis in
Czech was presented in [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]. This work provides an
annotated corpus of 1244 sentences from the restaurant reviews
domain and a baseline ABSA model. Hercig et al. [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ]
extended the dataset from [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ], nearly doubling its size and
presented results using several unsupervised methods for
word meaning representation.
        </p>
        <p>
          The work in [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ] creates a dataset in the domain of IT
product reviews. This dataset contains 200 annotated
sentences and 2000 short segments, both annotated with
sentiment and marked aspect terms (targets) without any
categorization and sentiment toward the marked targets. Using
5-fold cross validation on the aspect term extraction task
(TE) they achieved 65.79% F-measure on the short
segments and 30.27% F-measure on the long segments.
2.2
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>Neural Networks and Sentiment Analysis</title>
        <p>
          First attempt to estimate sentiment using a neural network
was presented in [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ]. The authors propose using Active
Deep Networks which is a semi-supervised algorithm. The
network is based on Restricted Boltzmann Machines. The
approach is evaluated on several review datasets
containing an earlier version of the movie review dataset created
1For the English RT dataset and Czech Facebook dataset it can be
also called the sentence-level.
by Pang and Lee [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ]. It outperforms the state of the art
approaches on these datasets.
        </p>
        <p>
          Ghiassi et al. [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ] use Dynamic Artificial Network for
sentiment analysis of Tweets. The network uses n-gram
features and creates a Twitter-specific lexicon. The
approach is compared to Support Vector Machines classifier
and achieves better results.
        </p>
        <p>
          Socher et al. [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] utilizes a Recursive Neural Tensor
Network trained on the Stanford Sentiment Treebank (SST).
The network is tested on the binary sentiment
classification and on the fine-grained (continuous number from 0 to
1) sentiment polarity scale. It outperforms the state of the
art methods on both tasks.
        </p>
        <p>
          A Deep Convolutional Neural Network is utilized for
sentiment classification in [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ]. Classification accuracies of
48.3% (5 sentiment levels) and 85.7% (binary) on the the
SST dataset are achieved.
        </p>
        <p>
          Several papers propose more general neural networks
used for NLP tasks that are tested also on sentiment
datasets. One of such methods is presented in [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]. A
Convolutional Neural Network (CNN) architecture is
proposed and tested on several datasets such as Movie Review
(MR) dataset and SST. The tasks were sentiment
classification (binary or 5 sentiment levels), subjectivity
classification (subjective/objective) and question type
classification. It proved state-of-the-art performance on all datasets.
        </p>
        <p>
          In [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ] a Dynamic Convolutional Neural Network is
proposed. A concept of dynamic k-max pooling is used in
this network. It is tested on sentiment analysis and
question classification tasks.
        </p>
        <p>
          The authors of [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ] propose two CNNs for ontology
classification, sentiment analysis and single-label
document classification. Their networks are composed of 9
layers out of which 6 are convolutional layers and 3
fullyconnected layers with different numbers of hidden units
and frame sizes. They show that the proposed method
significantly outperforms the baseline approaches (bag of
words) on English and Chinese corpora.
3
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Data</title>
      <sec id="sec-3-1">
        <title>In this work we use two types of corpora:</title>
        <p>• Aspect-level for the ABSA task and
• Document-level for the sentiment polarity task.
The properties of these corpora are shown in Table 1.
The English Aspect-level datasets come from the SemEval
ABSA tasks. Although we show properties of the datasets
from previous years, we report results only on the latest
datasets from the SemEval 2016.</p>
        <p>We do not use the Czech IT product datasets because
of its small size and because no results for the sentiment
polarity task have been reported using these datasets so
far. The Czech Facebook dataset has a label for bipolar
sentiment which we discard, similarly to the original
publication.
For all experiments we use 10-fold cross validation in
cases where there are no designated test and train data
splits.
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>System</title>
      <p>The proposed sentiment classification system can be
divided into two modules. The first one serves for data
preprocessing and creates the data representation while the
second one performs the classification. The classification
module utilizes three different neural network
architectures. All networks use the same preprocessing.
4.1</p>
      <sec id="sec-4-1">
        <title>Data Preprocessing and Representation</title>
        <p>
          The importance of data preprocessing has been proven in
many NLP tasks. The first step in our preprocessing chain
is removing the accents similarly to [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] and converting
the text to lower case. This process may lead to loss of
some information but we include it due to the fact that the
data we use are collected from the Internet and therefore
it may contain grammatical errors, misspellings and could
be written either with or without accents. Finally, all
numbers are replaced with one common token. We also
perform stemming utilizing the High Precision Stemmer [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ].
        </p>
        <p>The input feature of the neural networks is a sequence
of words in the document represented using the one hot
encoding. A dictionary is first created from the training
set. It contains a specified number of most frequent words.
The words are then represented by their indexes in the
dictionary. The words that are not in the dictionary are
assigned a reserved "Out of dictionary" index. An important
issue is the variable length of classified sentences.
Therefore, we cut the longer ones and pad the shorter ones to a
fixed length. The padding token has also a reserved
index. We use dictionary size 20,000 in all experiments.
The sentence length was set to 50 in all experiments with
document-level sentiment. We set the sequence length to
11 in the aspect-level sentiment experiments.
4.2</p>
        <p>
          CNN 1
This network was proposed by Kim in [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]. It is a
modification of the architecture proposed in [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ]. The first layer is
the embedding one. It learns a word vector of fixed length
k for each word. We use k = 300 in all experiments. It
uses one convolutional layer which is composed of a set
of filters of size n × k which means that it is applied on
sequence of n words and the whole word vector (k is the
length of the word vector). The application of such filters
results in a set of feature maps (results after applying the
convolutional filters to the input matrix). Kim proposes to
use multiple filter sizes (n = 3, 4, 5) and utilizes 100
filters of each size. Rectified linear units (Relu) are used as
activation function, drop-out rate is set to 0.5 and the
minibatch size is 50. After this step, a max-over-time pooling is
applied on each feature map and thus the most significant
features are extracted. The selection of one most
important feature from each feature map is supposed to ensure
invariance to the sentence length. The max pooling layer
is followed by a fully connected softmax layer which
outputs the probability distribution over labels. There are four
approaches to the training of the embedding layer:
• 1) Word vectors trained from scratch (randomly
initialized)
• 2) Static word2vec [
          <xref ref-type="bibr" rid="ref26">26</xref>
          ] vectors
• 3) Non-static vectors (initialized by word2vec and
then fine tuned)
• 4) Multichannel (both random initialized and
pretrained by word2vec).
        </p>
        <p>The hyper-parameters of the network was set on the
development set from SST-2 dataset2. We use identical
configuration in our experiments to allow comparability. We
implemented only the basic – randomly initialized version
of word embeddings. Figure 1 depicts the architecture of
the network.</p>
        <p>
          2Stanford Sentiment Treebank with neutral reviews removed and
binary labels
The architecture of this network was designed according
to [
          <xref ref-type="bibr" rid="ref27">27</xref>
          ] where it is successfully used for multi-label
document classification.
        </p>
        <p>
          Contrary to the work of Kim [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ] this network uses just
one size of the convolutional kernels and not the
combination of several sizes. The kernels have only 1 dimension
(1D) while Kim have used larger 2 dimensional kernels.
It was proven on the document classification task that the
simple 1D kernels give better results than the 2D ones.
        </p>
        <p>The input of the network is a vector of word indexes as
described in Section 4.1. The first layer is an embedding
layer which represents each input word as a vector of a
given length. The document is thus represented as a matrix
with l rows and k columns where k is the length of
embedding vectors. The embedding length is set to 300. The next
layer is the convolutional one. We use nc convolution
kernels of the size lk × 1 which means we do 1D convolution
over one position in the embedding vector over lk input
words. The size k is set to 3 (aspect-level sentiment) and 5
(document-level sentiment) in our experiments and we use
nc = 32 kernels. The following layer performs max
pooling over the length l − lk + 1 resulting in nn 1 × k vectors.
The output of this layer is then flattened and connected
with the output layer containing either 2 or 3 nodes
(number of sentiment labels). Figure 2 shows the architecture
of the network.
4.4</p>
      </sec>
      <sec id="sec-4-2">
        <title>LSTM</title>
        <p>The word sequence is the input to an embedding layer
same as for the CNNs. We use the embedding length of
300 in all experiments. The word embeddings are then
fed to the recurrent LSTM layer with 128 hidden neurons.
Dropout rate of 0.5 is then applied and the final state of the
LSTM layer is connected with the softmax output layer.
The network architecture is depicted in Figure 3.
4.5</p>
      </sec>
      <sec id="sec-4-3">
        <title>Tools</title>
        <p>
          We used Keras [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ] for implementation of all above
mentioned neural networks. It is based on the Theano deep
learning library [
          <xref ref-type="bibr" rid="ref29">29</xref>
          ]. It has been chosen mainly because
of good performance and our previous experience with this
tool. All experiments were computed on GPU to achieve
reasonable computation times.
5
        </p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Experiments</title>
      <p>
        Results on RT movie dataset [
        <xref ref-type="bibr" rid="ref30">30</xref>
        ] (10662 sentences, 2
classes) confirm that our implementation works similarly
to the original (see Table 2).
      </p>
      <p>We further performed evaluation on the current
SemEval 2016 ABSA dataset to allow comparison with the
current state-of-the-art methods. These results (see
Table 3) show that the used neural network architectures are
still quite far from the finely tuned state-of-the-art results.
However we need to remind the reader that our goal was
not to achieve the state-of-the-art results, but to replicate
network architectures that are used for sentiment analysis
in English as well as some networks utilized for other tasks
in Czech.</p>
      <p>
        Results on the Czech document-level datasets are shown
in Table 4. For the CSFD movie dataset, results are much
worse than the previous work. We believe that this is due
to the number of words used for representation. We used
50 words in all experiments and it may not suffice to fully
understand the review. This is supported by the fact that
the global target context [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] helps to improve the results
by 1.5%.
      </p>
      <p>
        We applied three types of neural networks to the term
polarity (TP) and class polarity (CP) tasks and evaluated
them on the Czech aspect-level restaurant reviews dataset.
The results in Table 5 show markedly inferior results
compared to the state-of-the-art results 72.5% for the TP and
75.2% for the CP tasks in [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]. Best results are achieved
using the combination of words and stemms as input.
      </p>
      <p>
        The inputs of the networks are one-hot vectors created
from words in the context window of the given aspect
term. We used five words in each direction of the searched
aspect term resulting in window size 11. We do not use any
weighting to give more importance to the closest words as
in [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ].
      </p>
      <p>For statistical significance testing, we report confidence
intervals at α 0.05.</p>
      <p>CNN1 and CNN2 present similar results although the
average best performance is achieved by the CNN2
architecture. The LSTM architecture consistently
underperforms, we believe that this is due to the basic architecture
model.
6</p>
    </sec>
    <sec id="sec-6">
      <title>Conclusion and Future Work</title>
      <p>In this work we have presented the first attempts to classify
sentiment of Czech sentences using a neural network. We
evaluated three architectures.</p>
      <p>We first performed experiments on two English corpora
mainly to allow comparability with existing work for
sentiment analysis in English.</p>
      <p>We have further experimented with three Czech corpora
for document-level sentiment analysis and one corpus for
aspect based sentiment analysis. The experiments proved
that the tested networks don’t achieve as good results as
the state-of-the-art approaches. The most promising
results were obtained when using the CNN2 architecture.
However, regarding the confidence intervals, we can
consider the performance of the architectures rather
comparable.</p>
      <p>The results show that Czech is much more complicated
to handle when determining sentiment polarity. This can
be caused by various properties of Czech language that
differ from English (e.g. double negative, sentence length,
comparative and superlative adjectives, or free word
order). Double or multiple negatives are grammatically
correct ways to express negation in Czech while in English
double negative is not acceptable in formal situations or
in writing. Thus the semantic meaning of sentences with
double or multiple negatives is hard to determine. In
English comparative and superlative forms of adjectives are
created by adding suffixes3 while in Czech suffixes and
prefixes are used. Informal texts can contain mixed
irregular adjectives with prefixes and/or suffixes thus making it</p>
      <sec id="sec-6-1">
        <title>3excluding irregular and long adjectives</title>
        <p>harder to determine the semantic meaning of these texts.
The free word order can also cause difficulties to train the
models because the same thing may be expressed
differently.</p>
        <p>However, it must be noted that the compared approaches
utilize much richer information than our basic features fed
to the neural networks. The neural networks were also not
fine-tuned for the task. Therefore we believe that there
is much room for further improvement and that neural
networks can reach or even outperform the state-of-the-art
results.</p>
        <p>We consider this paper to be the initial work on
sentiment analysis in Czech using neural networks. Therefore,
there are numerous possibilities for the future work. The
obtained results must be thoroughly analysed to identify
cases where the neural networks fail. An interesting
experiment would be sentiment analysis on Czech data
automatically translated to English. One possible direction
of further improvement is utilizing word embeddings to
initialize the embedding layer. We also plan to
experiment with neural networks on the other two tasks of aspect
based sentiment analysis – aspect term extraction and
aspect category extraction. Another perspective is to develop
new neural network architectures for sentiment analysis.</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgements</title>
      <p>This work was supported by the project LO1506 of the
Czech Ministry of Education, Youth and Sports and by
Grant No. SGS-2016-018 Data and Software
Engineering for Advanced Applications. Computational resources
were provided by the CESNET LM2015042 and the
CERIT Scientific Cloud LM2015085, provided under the
programme "Projects of Large Research, Development,
and Innovations Infrastructures".</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>Convolutional neural networks for sentence classification</article-title>
          .
          <source>arXiv preprint arXiv:1408.5882</source>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Socher</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Perelygin</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>J.Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chuang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manning</surname>
            ,
            <given-names>C.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ng</surname>
            ,
            <given-names>A.Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Potts</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Recursive deep models for semantic compositionality over a sentiment treebank</article-title>
          .
          <source>In: Proceedings of the conference on empirical methods in natural language processing (EMNLP)</source>
          . Volume
          <volume>1631</volume>
          .,
          <string-name>
            <surname>Citeseer</surname>
          </string-name>
          (
          <year>2013</year>
          )
          <fpage>1642</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>dos</given-names>
            <surname>Santos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.N.</given-names>
            ,
            <surname>Gatti</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          :
          <article-title>Deep convolutional neural networks for sentiment analysis of short texts</article-title>
          . In: COLING. (
          <year>2014</year>
          )
          <fpage>69</fpage>
          -
          <lpage>78</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Pontiki</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Galanis</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pavlopoulos</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Papageorgiou</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Androutsopoulos</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manandhar</surname>
            ,
            <given-names>S.:</given-names>
          </string-name>
          <article-title>SemEval-2014 Task 4: Aspect based sentiment analysis</article-title>
          .
          <source>In: Proceedings of the 8th International Workshop on Semantic Evaluation (SemEval</source>
          <year>2014</year>
          ), Dublin, Ireland, Association for Computational Linguistics and Dublin City University (
          <year>August 2014</year>
          )
          <fpage>27</fpage>
          -
          <lpage>35</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Pontiki</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Galanis</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Papageorgiou</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manandhar</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Androutsopoulos</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Semeval-2015 task 12: Aspect based sentiment analysis</article-title>
          .
          <source>In: Proceedings of the 9th International Workshop on Semantic Evaluation (SemEval</source>
          <year>2015</year>
          ),
          <article-title>Association for Computational Linguistics</article-title>
          , Denver, Colorado. (
          <year>2015</year>
          )
          <fpage>486</fpage>
          -
          <lpage>495</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Pontiki</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Galanis</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Papageorgiou</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Androutsopoulos</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manandhar</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>AL-Smadi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Al-Ayyoub</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhao</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Qin</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Clercq</surname>
            ,
            <given-names>O.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hoste</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Apidianaki</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tannier</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Loukachevitch</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kotelnikov</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bel</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jiménez-Zafra</surname>
            ,
            <given-names>S.M.</given-names>
          </string-name>
          , Eryig˘it, G.:
          <article-title>SemEval-2016 task 5: Aspect based sentiment analysis</article-title>
          .
          <source>In: Proceedings of the 10th International Workshop on Semantic Evaluation. SemEval '16</source>
          , San Diego, California, Association for Computational Linguistics (
          <year>June 2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Veselovská</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <article-title>Hajicˇ jr</article-title>
          ., J.,
          <string-name>
            <surname>Šindlerová</surname>
          </string-name>
          , J.:
          <article-title>Creating annotated resources for polarity classification in Czech</article-title>
          . In Jancsary, J., ed.
          <source>: Proceedings of KONVENS</source>
          <year>2012</year>
          , ÖGAI (
          <year>September 2012</year>
          )
          <fpage>296</fpage>
          -
          <lpage>304</lpage>
          PATHOS 2012 Workshop.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Steinberger</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lenkova</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kabadjov</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Steinberger</surname>
          </string-name>
          , R., van der Goot, E.:
          <article-title>Multilingual entity-centered sentiment analysis evaluated by parallel corpora</article-title>
          .
          <source>In: Proceedings of the 8th International Conference Recent Advances in Natural Language Processing</source>
          . (
          <year>2011</year>
          )
          <fpage>770</fpage>
          -
          <lpage>775</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Steinberger</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ebrahim</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            ,
            <surname>Hurriyetoglu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Kabadjov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Lenkova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Steinberger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            ,
            <surname>Tanev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            ,
            <surname>Vázquez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            ,
            <surname>Zavarella</surname>
          </string-name>
          ,
          <string-name>
            <surname>V.</surname>
          </string-name>
          :
          <article-title>Creating sentiment dictionaries via triangulation</article-title>
          .
          <source>Decision Support Systems</source>
          <volume>53</volume>
          (
          <year>2012</year>
          )
          <fpage>689</fpage>
          --
          <lpage>694</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Veselovská</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <article-title>Hajicˇ jr</article-title>
          ., J.:
          <article-title>Why words alone are not enough: Error analysis of lexicon-based polarity classifier for Czech</article-title>
          .
          <source>In: Proceedings of the 6th International Joint Conference on Natural Language Processing</source>
          , Nagoya, Japan,
          <source>Asian Federation of Natural Language Processing, Asian Federation of Natural Language Processing</source>
          (
          <year>2013</year>
          )
          <fpage>1</fpage>
          -
          <lpage>5</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Habernal</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          , Ptácˇek, T.,
          <string-name>
            <surname>Steinberger</surname>
          </string-name>
          , J.:
          <article-title>Sentiment analysis in Czech social media using supervised machine learning</article-title>
          .
          <source>In: Proceedings of the 4th Workshop on Computational Approaches</source>
          to Subjectivity, Sentiment and
          <string-name>
            <surname>Social Media Analysis</surname>
          </string-name>
          , Atlanta,
          <string-name>
            <surname>GA</surname>
          </string-name>
          , USA, Association for Computational Linguistics (
          <year>June 2013</year>
          )
          <fpage>65</fpage>
          -
          <lpage>74</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Habernal</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          , Ptácˇek, T.,
          <string-name>
            <surname>Steinberger</surname>
          </string-name>
          , J.:
          <article-title>Supervised sentiment analysis in Czech social media</article-title>
          .
          <source>Information Processing &amp; Management</source>
          <volume>50</volume>
          (
          <issue>5</issue>
          ) (
          <year>2014</year>
          )
          <fpage>693</fpage>
          -
          <lpage>707</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Habernal</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brychcín</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Semantic spaces for sentiment analysis</article-title>
          .
          <source>In: Text, Speech and Dialogue</source>
          . Volume
          <volume>8082</volume>
          of Lecture Notes in Computer Science., Berlin, SpringerVerlag (
          <year>2013</year>
          )
          <fpage>482</fpage>
          -
          <lpage>489</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Brychcín</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Konopík</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Semantic spaces for improving language modeling</article-title>
          .
          <source>Computer Speech &amp; Language</source>
          <volume>28</volume>
          (
          <issue>1</issue>
          ) (
          <year>2014</year>
          )
          <fpage>192</fpage>
          -
          <lpage>209</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Brychcín</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Habernal</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Unsupervised improving of sentiment analysis using global target context</article-title>
          .
          <source>In: Proceedings of the International Conference Recent Advances in Natural Language Processing RANLP</source>
          <year>2013</year>
          , Shoumen, Bulgaria, INCOMA Ltd. (
          <year>September 2013</year>
          )
          <fpage>122</fpage>
          -
          <lpage>128</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Steinberger</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brychcín</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Konkol</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Aspect-level sentiment analysis in Czech</article-title>
          .
          <source>In: Proceedings of the 5th Workshop on Computational Approaches</source>
          to Subjectivity, Sentiment and
          <string-name>
            <surname>Social Media Analysis</surname>
          </string-name>
          , Baltimore,
          <string-name>
            <surname>MD</surname>
          </string-name>
          , USA, Association for Computational Linguistics (
          <year>June 2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Hercig</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brychcín</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Svoboda</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Konkol</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Steinberger</surname>
          </string-name>
          , J.:
          <article-title>Unsupervised methods to improve aspect-based sentiment analysis in Czech. Computación y Sistemas (in press)</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Tamchyna</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fiala</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Veselovská</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>Czech aspectbased sentiment analysis: A new dataset and preliminary results</article-title>
          .
          <source>Proceedings of the 15th conference ITAT</source>
          <year>2015</year>
          (
          <year>2015</year>
          )
          <fpage>95</fpage>
          -
          <lpage>99</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Zhou</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          :
          <article-title>Active deep networks for semi-supervised sentiment classification</article-title>
          .
          <source>In: Proceedings of the 23rd International Conference on Computational Linguistics: Posters</source>
          , Association for Computational Linguistics (
          <year>2010</year>
          )
          <fpage>1515</fpage>
          -
          <lpage>1523</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>Pang</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vaithyanathan</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Thumbs up?: sentiment classification using machine learning techniques</article-title>
          .
          <source>In: Proceedings of the ACL-02 conference on Empirical methods in natural language processing-</source>
          Volume
          <volume>10</volume>
          ,
          <string-name>
            <surname>Association for Computational Linguistics</surname>
          </string-name>
          (
          <year>2002</year>
          )
          <fpage>79</fpage>
          -
          <lpage>86</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <surname>Ghiassi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Skinner</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zimbra</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Twitter brand sentiment analysis: A hybrid system using n-gram analysis and dynamic artificial neural network</article-title>
          .
          <source>Expert Systems with applications</source>
          <volume>40</volume>
          (
          <issue>16</issue>
          ) (
          <year>2013</year>
          )
          <fpage>6266</fpage>
          -
          <lpage>6282</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <surname>Kalchbrenner</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grefenstette</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Blunsom</surname>
            ,
            <given-names>P.:</given-names>
          </string-name>
          <article-title>A convolutional neural network for modelling sentences</article-title>
          .
          <source>arXiv preprint arXiv:1404.2188</source>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>LeCun</surname>
          </string-name>
          , Y.:
          <article-title>Text understanding from scratch</article-title>
          .
          <source>arXiv preprint arXiv:1502.01710</source>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <surname>Brychcín</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Konopík</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Hps: High precision stemmer</article-title>
          .
          <source>Information Processing &amp; Management</source>
          <volume>51</volume>
          (
          <issue>1</issue>
          ) (
          <year>2015</year>
          )
          <fpage>68</fpage>
          -
          <lpage>91</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <surname>Collobert</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Weston</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bottou</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Karlen</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kavukcuoglu</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kuksa</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Natural language processing (almost) from scratch</article-title>
          .
          <source>The Journal of Machine Learning Research</source>
          <volume>12</volume>
          (
          <year>2011</year>
          )
          <fpage>2493</fpage>
          -
          <lpage>2537</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Corrado</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dean</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          :
          <article-title>Efficient estimation of word representations in vector space</article-title>
          . (
          <year>2013</year>
          ) arXiv preprint arXiv:
          <volume>1301</volume>
          .
          <fpage>3781</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <surname>Lenc</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Král</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Deep neural networks for Czech multilabel document classification</article-title>
          .
          <source>In: International Conference on Intelligent Text Processing and Computational Linguistics</source>
          , Konya,
          <source>Turkey (April 3 - 9</source>
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <surname>Chollet</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          : Keras. https://github.com/fchollet/ keras (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <surname>Bergstra</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Breuleux</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bastien</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lamblin</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pascanu</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Desjardins</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Turian</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Warde-Farley</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bengio</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>Theano: a cpu and gpu math expression compiler</article-title>
          .
          <source>In: Proceedings of the Python for scientific computing conference (SciPy)</source>
          . Volume
          <volume>4</volume>
          ., Austin, TX (
          <year>2010</year>
          )
          <fpage>3</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <surname>Pang</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Seeing stars: Exploiting class relationships for sentiment categorization with respect to rating scales</article-title>
          .
          <source>In: Proceedings of the ACL</source>
          . (
          <year>2005</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>