<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Czech Aspect-Based Sentiment Analysis: A New Dataset and Preliminary Results</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Aleš Tamchyna</string-name>
          <email>tamchyna@ufal.mff.cuni.cz</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ondrˇej Fiala</string-name>
          <email>fiala@ufal.mff.cuni.cz</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Katerˇina Veselovská</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Charles University in Prague, Faculty of Mathematics and Physics Institute of Formal and Applied Linguistics Malostranské námeˇstí 25</institution>
          ,
          <addr-line>Prague</addr-line>
          ,
          <country country="CZ">Czech Republic</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2015</year>
      </pub-date>
      <fpage>95</fpage>
      <lpage>99</lpage>
      <abstract>
        <p>This work focuses on aspect-based sentiment analysis, a relatively recent task in natural language processing. We present a new dataset for Czech aspect-based sentiment analysis which consists of segments from user reviews of IT products. We also describe our work in progress on the task of aspect term extraction. We believe that this area can be of interest to other workshop participants and that this paper can inspire a fruitful discussion on the topic with researchers from related fields.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Sentiment analysis (or opinion mining) is a field related
to natural language processing (NLP) which studies how
people express emotions (or opinions, sentiments,
evaluations) in language and which develops methods to
automatically identify such opinions.</p>
      <p>The most typical task of sentiment analysis is to look at
some short text (a sentence, paragraph, short review) and
determine its polarity – positive, negative or neutral.</p>
      <p>Aspect-based sentiment analysis (ABSA) refers to
discovering aspects (aspect terms, opinion targets) in text and
classifying their polarity. The prototypical scenario are
product reviews: we assume that products have several
aspects (such as size or battery life for cellphones) and we
attempt to identify users’ opinions on these individual
aspects.</p>
      <p>This is a more fine-grained approach than the standard
formulation of sentiment analysis where the goal would be
to classify the polarity of entire sentences (or even whole
reviews) without regard for internal structure.</p>
      <p>
        Recently, ABSA has been gaining researchers’ interest,
as evidenced e.g. by the two consecutive shared tasks
organized within SemEval in 2014 and 2015 [
        <xref ref-type="bibr" rid="ref6 ref7">7, 6</xref>
        ].
      </p>
      <p>ABSA can be roughly divided into two subtasks:
(i) identification of aspects (or aspect term extraction) in
text, i.e. marking (occurrences of) words which are
evaluated; (ii) polarity classification, i.e. deciding whether the
opinions about the identified words are positive, negative
or neutral.</p>
      <p>In this work, we introduce a new Czech dataset of
product reviews annotated for ABSA and describe a
preliminary method of aspect term identification which combines
a rule-based approach and machine learning.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Dataset of IT Product Reviews</title>
      <p>We downloaded a number of user product reviews which
are publicly available on the website of an established
Czech online shop with electronic devices. Each review
consists of negative and positive aspects of the product.
This setting pushes the customer to rate its important
characteristics.</p>
      <p>The dataset consists of two parts: (i) random short
segments and (ii) longest reviews. The difference in length is
reflected also in the use of language.</p>
      <p>The first part of this dataset contains 1000 positive and
1000 negative reviews which were selected from source
data and their targets were manually tagged. These
targets were either aspects of the evaluated product or some
general attributes (e.g. price, ease of use). The polarity
of each aspect is based on whether the user submitted the
segment as negative or positive. These short reviews often
contain only the aspect without any evaluative phrase.</p>
      <p>The second part of dataset consists of the longest
reviews. We chose 100 of them for each polarity.
These reviews represent more usual text and they tend to
keep proper sentence structure. The longest review has
7057 characters.</p>
      <p>The whole dataset provides a consistent view of
language used in the on-line environment preserving both
specific word forms and language structures. There is also
a large amount of domain specific slang due to the origin
of the text.</p>
      <sec id="sec-2-1">
        <title>Dataset part</title>
        <p>Random, positive
Random, negative
Longest, positive
Longest, negative
#targets
640
508
484
353
#reviews
1000
1000
100
100</p>
      </sec>
      <sec id="sec-2-2">
        <title>Avg. length</title>
        <p>34.17
39.72
953.35
855.04</p>
        <p>The data was annotated by a single annotator. The basic
instruction was to mark all aspects or general
characteristics of the product. The span of the annotated term should
be as small as possible (often a single noun). For
evaluation, the span can be expanded e.g. to the immediate
dependency subtree of the target. Any part of speech can
be marked; e.g. both “funk cˇnost” (“functionality”) and
“funk cˇní” (“functional”) should be marked.
a-tree
zone=en</p>
        <p>The fried
AuxA Atr
DT JJ
is .</p>
        <p>Pred AuxK</p>
        <p>VBZ .
rice amazing
Sb Pnom</p>
        <p>NN JJ</p>
        <p>The whole dataset contains 1985 target tags; 1124 of
these are positive and 861 are negative. Detailed target
statistics are shown in Table 1.</p>
        <p>The dataset is freely available for download at the
following URL:</p>
        <p>http://hdl.handle.net/11234/1-1507.
3</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Pipeline</title>
      <p>
        Our work is inspired by the pipeline of [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. We run
morphological analysis and tagging on the data to identify the
parts of speech of words and their morphological features
(e.g. case or gender for Czech). We also obtain
dependency parses of the sentences. Then, we use several
handcrafted rules based on syntax to mark the likely aspects in
the data. Figure 1 shows a sample dependency parse tree
and rule application.
      </p>
      <p>
        Unlike [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], the core of our approach is a
machinelearning model and the outputs of the rules only serve as
additional “hints” (features) to help the model identify
aspects.
3.1
      </p>
      <sec id="sec-3-1">
        <title>Syntactic Rules</title>
        <p>
          We use the same rules as [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ], Table 2 contains their
description. Here, we categorize the rules somewhat
differently, their types correspond to the actual features
presented to the model.
        </p>
        <p>The rules are designed for opinion target identification,
i.e. discovering targets of evaluative statements.1 They
are based on syntactic relations with evaluative words, i.e.</p>
        <p>1The underlying assumption of this approach is that opinion targets
tend to be the sought-after aspects.
words listed in a subjectivity lexicon for the given
language.</p>
        <p>In the example in Figure 1, the rule vbnm_sb_adj is
triggered because “amazing” is an evaluative word and it
is a predicate adjective – the word “rice”, as the subject
of this syntactic construction, is then marked as a likely
aspect term.</p>
        <p>
          Originally, the rules were written for English. Their
adaptation to Czech proved very simple. We modified
expressions which involved morphological tags to work with
the Czech positional tagset [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]. Some of the rules included
lexical items, such as the lemma “be” for identifying the
linking verbs of predicate nominals. Simple translation of
these few words to Czech sufficed in such cases.
3.2
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>Model</title>
        <p>
          We chose linear-chain conditional random fields (CRFs)
for our work [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]. In this model, aspect identification is
viewed as a sequence labeling task. The input x are words
in the sentence and the output is a labeling y of the same
length: each word is marked as either the beginning of an
aspect (B), inside an aspect (I) or outside an aspect (O).2
        </p>
        <p>A linear-chain CRF is a statistical model. It is related to
hidden Markov models (HMMs), however it is a
discriminative model, not a generative one – it directly models
the conditional probability of the labeling P(y|x).
Linearchain CRFs assume that the probability of the current label
(B, I or O) only depends on the previous label and on the
input words x.</p>
        <p>Formally, a linear-chain CRF is the following
conditional probability distribution:</p>
        <p>P(y|x) =</p>
        <p>1 T K
Z(x) exp{t∑=1 k∑=1 λk fk(yt , yt−1, t, x)}
(1)</p>
        <p>Roughly speaking, P(y|x) is the score of the sentence
labeling y, exponentiated and normalized.</p>
        <p>The score of y corresponds to the sum of scores for
labels yt at each position t ∈ {1, . . . , T } in the sentence.
The score at position t is the product between the values
of feature functions fk(yt , yt−1, t, x) and their associated
weights λk, which are estimated in the learning stage.</p>
        <p>Feature functions can look at the current label yt , the
previous label yt−1 and the whole input sentence x (which
is constant).</p>
        <p>Z(x) is the normalization function which sums over all
possible label sequences:</p>
        <p>T K
Z(x) = ∑ exp{∑ ∑ λk fk(yt0 , yt0−1, t, x)}
y0 t=1 k=1
(2)</p>
        <p>To train the model, we require training data, i.e.
sentences with the labeling already assigned by a human
annotator. During CRF learning, the weights λk are
optimized to maximize the likelihood of the observed labeling
2This “BIO” labeling scheme is common for CRFs. In practice, it
brings us a consistent slight improvement as opposed to using only binary
classification (inside vs. outside an aspect).
ID
adverb
but_opposite
coord
sub_adj
subj_of_pat
verb_actant_pat
verb_actant_act
vbnm_patn
vbnm_sb_adj</p>
        <sec id="sec-3-2-1">
          <title>Description</title>
          <p>Actor or patient of a verb with a subjective adverb.
Words coordinated with an aspect with “but”.
Words coordinated with an aspect are also aspects.
Nouns modified by subjective adjectives.
Subject of a clause with a subjective patient.
Patient of a transitive evaluative verb.</p>
          <p>Actor of an intransitive evaluative verb.</p>
          <p>Predicative nominal (patient).</p>
          <p>Subject of predicative adjectives.</p>
        </sec>
        <sec id="sec-3-2-2">
          <title>Example</title>
          <p>The pizza tastes so good.</p>
          <p>The food is outstanding, but everything else sucks.</p>
          <p>The excellent mussels, goat cheese and salad.</p>
          <p>A very capable kitchen.</p>
          <p>The bagel have an outstanding taste.</p>
          <p>I liked the beer selection.</p>
          <p>Their wine sucks.</p>
          <p>Our favourite meal is the sausage.</p>
          <p>The fried rice is amazing.
in the dataset. Gradient-based optimization techniques are
usually applied for learning.</p>
          <p>At prediction time, the weights λk are fixed and we are
looking for such a labeling yˆ which is the most probable
according to the model, i.e.:
yˆ = arg max P(y|x)
y
(3)
yˆ can be found efficiently using a variant of the Viterbi
algorithm (dynamic programming). In our work, we use
the CRF++ toolkit3 both for training and prediction.
3.3</p>
        </sec>
      </sec>
      <sec id="sec-3-3">
        <title>Feature Set</title>
        <p>We now describe the various feature sets evaluated in this
work.</p>
        <p>Surface features. We use the surface forms of the
current word, two preceding and two following words as
separate features. Additionally, we extract all (four) bigrams
and (three) trigrams of surface forms from this window.
We also use the CRF++ bigram feature template without
any arguments; this simply produces the concatenation of
the previous and current label (yt−1, yt ).</p>
        <p>Morpho-syntactic features. We extract unigrams,
bigrams and trigrams from a limited context window
(identical to the above) around the current token but instead of
surface forms, we look at:
• lemma,
• morphological tag,
• analytical function.</p>
        <p>Analytical functions are assigned by the dependency
parser and their values include “Sb” for subject, “Pred”
for predicate etc.</p>
        <p>Sublex features. We mark all words in the data whose
lemma is found in the subjectivity lexicon. For each
token in the window of size 4 around the current token
(included), we extract a feature indicating whether it was
marked as subjective. We also concatenate these
indicator features with the surface form of the current token.
3http://taku910.github.io/crfpp/</p>
        <p>Rule features. Finally, for each type of rule, we
extract features for the current token, the preceding and the
following token, indicating whether the rule marked that
token. Again, these features have two versions: one
standalone and one concatenated with the surface form of the
current token.
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Experiments</title>
      <p>
        We analyze our data using Treex [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], a modular NLP
toolkit. Sentences are first tokenized and tagged using
Morphodita [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. Then we obtain their dependency parses
using the MST parser [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. We use Czech SubLex [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] is
our subjectivity lexicon both for the CRF sublex features
and for the rules. The rules are implemented as blocks
within the Treex platform.
4.1
      </p>
      <sec id="sec-4-1">
        <title>Results</title>
        <p>Table 3 shows the obtained precision (P), recall (R) and
f-measure (F1) for both parts of the data set. The results
in all cases were acquired using 5-fold cross-validation on
the training data.</p>
        <p>Random segments. The baseline (surface-only)
features achieve the best precision but the recall is very low.</p>
        <p>Morpho-syntactic features lower the precision by a
significant margin but push recall considerably. As the review
data come from the “wild”, they are quite noisy; many
segments are written without punctuation, reducing the
benefit of using morphological analysis, let alone dependency
parsing.4</p>
        <p>Often, the segments are rather short, such as “Rychlé
dodání” (“fast delivery”) or “ Fotky fakt parádní.” (“
Photos really awesome.”). This also considerably limits the
benefit that a parser can bring – there is a major domain
mismatch both in the text topic and types of sentences
between the parser’s training data and this dataset, so we
cannot expect parsing accuracy to be high.</p>
        <p>Most of the improvement from adding
morphosyntactic features thus probably comes from the
availability of word lemmas – this allows the CRF to learn which
4This issue could perhaps be addressed by using a spell-checker, we
leave that to future work.</p>
      </sec>
      <sec id="sec-4-2">
        <title>Feature set</title>
        <p>surface
+morpho-syntactic
+sublex
+rules
words are frequently marked as aspects in this domain
and to generalize this information beyond their current
inflected form.</p>
        <p>Adding the information from the sentiment lexicon
further improves performance, though not as much as we
would expect. We could possibly further increase its
impact through more careful feature engineering – so far, the
features only capture whether a subjective term is present
in a small linear context. For example, the lemma of the
evaluative word could be included in the feature.5</p>
        <p>Finally, adding the output of syntactic rules further
improves the results. Due to the uncommon syntactic
structure of the segments, most rules were not active very often,
so the space for improvement is quite limited. Yet the
results show that when the rules do trigger, their output can
be a useful signal for the CRF.</p>
        <p>
          The observed improvement in recall at the slight
expense of precision is in line with the results of [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] where
the system based on the same rules achieved high recall
and rather low precision.
        </p>
        <p>Long reviews. It is immediately apparent that the long
reviews are a much more difficult dataset than review
segments – the best f-measure achieved on the short segments
is 65.79 while here it is only 30.27. This can be explained
by the lower density of aspect terms compared to random
review segments and a much higher sentence length –
after sentence segmentation, the average sentence length is
over 29 words, compared to only 6 words for the random
segments.</p>
        <p>When using only the baseline features, the recall is
extremely low. Adding morpho-syntactic features has a
similar effect as for the random segments – precision is
lowered but recall nearly triples.</p>
        <p>Interestingly, adding features from the subjectivity
lexicon changes the picture considerably. This feature set
obtains the highest precision but recall is lower compared to
both +morpho-syntactic and +rules. It may be that due to
the high sentence length, sublex features help identify
aspects within the short window but their presence pushes
the model to ignore the more distant ones. A more
thorough manual evaluation would be required to confirm this.</p>
        <p>Finally, the addition of syntactic rules leads to the
highest f-measure, even though neither recall nor precision are
the best. In this dataset, possibly again thanks to the length
5CRF++ feature templates do not offer a simple way to achieve this
without also generating a large number of uninformative feature types.
of sentences, the rules are trigged much more often than
for the random segments. Rule features can therefore have
a more prominent effect on the model.
5</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Related Work</title>
      <p>
        In terms of using rules for ABSA, our work is inspired
by [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. Such rules can also be used iteratively to
expand both the aspects and evaluative terms using the
double propagation algorithm [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. Other methods of
discovering opinion targets are described, inter alia, in [
        <xref ref-type="bibr" rid="ref3 ref5 ref9">3, 9, 5</xref>
        ].
Linear-chain CRFs have been applied in sentiment
analysis and they are also well suited for ABSA, they were used
e.g. by the winning submission by [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] to the SemEval
2014 Task 4.
      </p>
      <p>
        For Czech, a dataset for ABSA was published by [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
This dataset is in the domain of restaurant reviews and
closely follows the methodology of [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Our work focuses
on reviews of IT products, naturally complementing this
dataset. It should further support research in this area and
enable researchers to evaluate their approaches on diverse
domains.
6
      </p>
    </sec>
    <sec id="sec-6">
      <title>Conclusion</title>
      <p>We have presented a new dataset for ABSA in the Czech
language and we have described a baseline system for the
subtask of aspect term extraction.</p>
      <p>The dataset consists of segments from user reviews of
IT products with the annotation of aspects and their
polarity.</p>
      <p>The system for aspect term extraction is based on
linear-chain CRFs and uses a number of surface and
linguistically-informed features. On top of these features,
we have shown that task-specific syntactic rules can
provide useful input to the model.</p>
      <p>Utility of the syntactic rules could be further evaluated
on other domains (such as the Czech restaurant reviews)
or languages (e.g. using the official SemEval data sets)
and the impact of individual rules could be thoroughly
analyzed across these data sets.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgements</title>
      <p>This research was supported by the grant GA15-06894S
of the Grant Agency of the Czech Republic and by the
SVV project number 260 224. This work has been
using language resources developed, stored and distributed
by the LINDAT/CLARIN project of the Ministry of
Education, Youth and Sports of the Czech Republic (project
LM2010013).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Hajicˇ</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vidová-Hladká</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Tagging inflective languages: Prediction of morphological categories for a rich, structured tagset</article-title>
          .
          <source>In: Proceedings of the COLING - ACL Conference</source>
          ,
          <volume>483</volume>
          -
          <fpage>490</fpage>
          ,
          <year>1998</year>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Lafferty</surname>
            ,
            <given-names>J. D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>McCallum</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pereira</surname>
            ,
            <given-names>F. C. N.</given-names>
          </string-name>
          :
          <article-title>Conditional random fields: Probabilistic models for segmenting and labeling sequence data</article-title>
          .
          <source>In: Proceedings of the Eighteenth International Conference on Machine Learning</source>
          , ICML'
          <volume>01</volume>
          ,
          <fpage>282</fpage>
          -
          <lpage>289</lpage>
          , San Francisco, CA, USA,
          <year>2001</year>
          , Morgan Kaufmann Publishers Inc.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Web data mining: exploring hyperlinks, contents, and usage data (Data-centric systems</article-title>
          and applications). Springer-Verlag New York, Inc., Secaucus, NJ, USA,
          <year>2006</year>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>McDonald</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pereira</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ribarov</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hajicˇ</surname>
          </string-name>
          . J.:
          <article-title>Nonprojective dependency parsing using spanning tree algorithms</article-title>
          .
          <source>In: Proceedings of the conference on Human Language Technology and Empirical Methods in Natural Language Processing</source>
          ,
          <fpage>523</fpage>
          -
          <lpage>530</lpage>
          ,
          <year>2005</year>
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Qiaozhu</given-names>
            <surname>Mei</surname>
          </string-name>
          , Xu Ling, Matthew Wondra, Hang Su, and ChengXiang Zhai.
          <article-title>Topic sentiment mixture: Modeling facets and opinions in weblogs</article-title>
          .
          <source>In Proceedings of the 16th International Conference on World Wide Web, WWW '07</source>
          , pages
          <fpage>171</fpage>
          -
          <lpage>180</lpage>
          , New York, NY, USA,
          <year>2007</year>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Pontiki</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Galanis</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Papageorgiou</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manandhar</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Androutsopoulos</surname>
          </string-name>
          , I.: SemEval-2015 task 12:
          <article-title>Aspect based sentiment analysis</article-title>
          .
          <source>In: Proceedings of the 9th International Workshop on Semantic Evaluation (SemEval</source>
          <year>2015</year>
          ),
          <fpage>486</fpage>
          -
          <lpage>495</lpage>
          , Denver, Colorado, June 2015, Association for Computational Linguistics
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Pontiki</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Galanis</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pavlopoulos</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Papageorgiou</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Androutsopoulos</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manandhar</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :. SemEval
          <article-title>-2014 task 4: Aspect based sentiment analysis</article-title>
          .
          <source>In: Proceedings of the 8th International Workshop on Semantic Evaluation (SemEval</source>
          <year>2014</year>
          ),
          <fpage>27</fpage>
          -
          <lpage>35</lpage>
          , Dublin, Ireland,
          <year>August 2014</year>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          and Dublin City University
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Popel</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Žabokrtský</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          :
          <article-title>TectoMT: modular NLP framework</article-title>
          . In: Hrafn Loftsson, Eirikur Rögnvaldsson, and Sigrun Helgadottir, (eds.),
          <source>IceTAL</source>
          <year>2010</year>
          , volume
          <volume>6233</volume>
          of Lecture Notes in Computer Science,
          <volume>293</volume>
          -
          <fpage>304</fpage>
          ,
          <article-title>Iceland Centre for Language Technology (ICLT</article-title>
          ), Springer,
          <year>2010</year>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Popescu</surname>
            ,
            <given-names>A. -M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Etzioni</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          :
          <article-title>Extracting product features and opinions from reviews</article-title>
          .
          <source>In: Proceedings of the Conference on Human Language Technology and Empirical Methods in Natural Language Processing</source>
          , HLT'
          <volume>05</volume>
          ,
          <fpage>339</fpage>
          -
          <lpage>346</lpage>
          , Stroudsburg, PA, USA,
          <year>2005</year>
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Qiu</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bu</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Opinion word expansion and target extraction through double propagation</article-title>
          .
          <source>Computational Linguistics</source>
          <volume>37</volume>
          (
          <issue>1</issue>
          ) (
          <year>March 2011</year>
          ),
          <fpage>9</fpage>
          -
          <lpage>27</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Steinberger</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brychcín</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Konkol</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Aspect-level sentiment analysis in Czech</article-title>
          .
          <source>In: Proceedings of the 5th Workshop on Computational Approaches</source>
          to Subjectivity, Sentiment and
          <string-name>
            <surname>Social Media Analysis</surname>
          </string-name>
          , Baltimore, USA,
          <year>June 2014</year>
          , Association for Computational Linguistics
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Straková</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Straka</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hajicˇ</surname>
          </string-name>
          , J.:
          <article-title>Open-source tools for morphology, lemmatization, POS tagging and named entity recognition</article-title>
          .
          <source>In: Proceedings of 52nd Annual Meeting of the Association for Computational Linguistics: System Demonstrations</source>
          ,
          <fpage>13</fpage>
          -
          <lpage>18</lpage>
          , Baltimore, Maryland, June 2014, Association for Computational Linguistics
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Toh</surname>
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>W.:</given-names>
          </string-name>
          <article-title>DLIREC: Aspect term extraction and term polarity classification system</article-title>
          .
          <source>In: Proceedings of the 8th International Workshop on Semantic Evaluation (SemEval</source>
          <year>2014</year>
          ),
          <fpage>235</fpage>
          -
          <lpage>240</lpage>
          , Dublin, Ireland,
          <year>August 2014</year>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          and Dublin City University
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Veselovská</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bojar</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          :
          <source>Czech SubLex 1.0</source>
          , 2013
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Veselovská</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tamchyna</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>ÚFAL: Using hand-crafted rules in aspect based sentiment analysis on parsed data</article-title>
          .
          <source>In: Proceedings of the Eighth International Workshop on Semantic Evaluation (SemEval</source>
          <year>2014</year>
          ),
          <fpage>694</fpage>
          -
          <lpage>698</lpage>
          , Dublin, Ireland,
          <year>2014</year>
          , Dublin City University, Dublin City University
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>