<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Di erent Aggregation Strategies for Generically Contextualized Sentiment Lexicons</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Stefan Gindl</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of New Media Technology, MODUL University Vienna</institution>
          ,
          <country country="AT">Austria</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Sentiment detection has gained relevance in the last years due to the vast amount of publicly available opinion in the form of Web forums or blogs. Yet, it still su ers from the ambiguity of language, lowering the e cacy and accuracy of sentiment detection systems. Thus, it is important to also invoke context information to re ne the initial values of sentiment terms. Moreover, domain-independence is desirable to avoid using a topic determination beforehand. This work investigates strategies for extracting non-generic features to be integrated into a socalled contextualized sentiment lexicon, capable of getting the context correctly and assigning sentiment terms the proper sentiment value. The proposed approach will be applied in an online-media aggregation and visualization portal, covering a vast number of news media sources.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>Sentiment detection handles a ect expressed in written text, more exactly it
tries to classify documents into positively, negatively or neutrally opinionated.
The classi cation can either be coarse-grained (i.e. positive, negative, neutral)
or ne-grained (i.e. strong-positive, weak-positive, etc.). The research area
experienced a leap in relevance with the upcoming availability of online opinions
in reviews, forums or blogs. Applications range from the political area
(tracking a political campaign online) over the economic area (acceptance studies for
new products or services) to the purely scienti c application, helping to
understand human language. Thus, sentiment detection can play a major role in Web
mining systems. It also adds value to Social Web applications. Trend analyses
on fast moving platforms such as www.twitter.com become possible; websites
hosting images or videos (such as www. ickr.com or www.youtube.com) can be
exploited to measure the a ect of the community towards celebrities or popular
technical devices.</p>
      <p>Many approaches rely on so-called sentiment lexicons, containing terms
assumed to express sentiment. Sentiment lexicons su er from term ambiguity - one
and the same term can have di erent meanings under di erent circumstances.
Table 1 shows three sentence, where one and the same sentiment term can be
used in positive and negative context. The intuitively negative term \repair" can
be used positively, when a person is satis ed with his/her repaired car.
\Unpredictable" applied to the movie genre refers to an exciting movie; on the other
hand, if the breaks of a car are unpredictable, this is normally something
undesirable. Finally, the term \peace" will be express a positive fact in the most
cases. Yet, it can also refer to a negative state, such as in the sentence \This
peace is a lie".</p>
      <sec id="sec-1-1">
        <title>Positive</title>
      </sec>
      <sec id="sec-1-2">
        <title>The repair of my car was satisfying.</title>
      </sec>
      <sec id="sec-1-3">
        <title>This movie's plot is unpredictable.</title>
      </sec>
      <sec id="sec-1-4">
        <title>The long peace brought wealth and safety to the people.</title>
      </sec>
      <sec id="sec-1-5">
        <title>Negative</title>
      </sec>
      <sec id="sec-1-6">
        <title>I had many complaints after my camera's repair.</title>
      </sec>
      <sec id="sec-1-7">
        <title>The breaks of this car are</title>
        <p>unpredictable.</p>
      </sec>
      <sec id="sec-1-8">
        <title>This peace is a lie. Table 1. Examples for sentiment terms occurring in positive and negative contexts.</title>
        <p>
          This work examines possible re nement strategies of the already existing
context-sensitive sentiment detection system described in [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]. It takes into
account the context of a sentiment term, and, based on the context, re nes the
sentiment value of the term. Nave Bayes as a simple, fast and yet powerful
technique serves as the method to train the model. To overcome the e ects of
domain-speci city the approach also merges features of the trained models and
creates a domain-independent model. In the presented paper re nement
strategies for creating a domain-independent lexicon are discussed, together with a
preliminary evaluation of the planned strategies.
        </p>
        <p>Temporal Sentiment Analysis Applied to Online Media
The proposed system will be used for temporal sentiment detection in the
socalled \Media Watch on Climate Change". This portal aggregates climate change
related issues and provides e cient visualization means, such as a semantic map
for related keywords with strong media coverage and an ontology map for
relations among signi cant phrases.</p>
        <p>The sentiment map in the upper left corner of Figure 1 allows for tracing
the sentiment towards relevant topics. For example, the phrases \oil spill" and
\gulf oil" receive clearly negative media attention, whereas the term \Hayward"
received positive attention until May 10, which turns into negative afterwards.
Such a tool, i.e. accurate sentiment detection combined with e cient
visualization techniques, strongly supports research on relevant topics and o ers a
specialized view on the online world.</p>
        <p>During the U.S. elections 2008 another portal website using a former
version of the proposed appraoch traced media attention towards the presidential
candidates. Figure 2 shows the main window of the portal, with the presidential
candidates in the upper part, a list of used media sources in the middle and the
sentiment map at the bottom. Such tools can complement or even replace
traditional opinion surveys, and are a permanent source of feedback during a political
campaign. Adapted to di erent application elds they can support enterpises to
trace their reputation (e.g. in connection with the current oil spill in the Gulf of
Mexico) or to measure the acceptance of a previously launched new product in
the online community.</p>
        <p>The paper is structured as follows: Section 2 summarizes existing work,
Section 3 outlines the already existing approach and the re nement strategies. The
evaluation follows in Section 4. Section 5 concludes the paper and contains an
outlook on further work regarding the discussed re nement strategies.
2</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>
        Sentiment detection as a research area dates back to the 1990s with the work
of Wiebe [20] and Hatzivassiloglou and McKeown [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. In [20] Wiebe started to
identify subjective sentences, whereas Hatzivassiloglou and McKeown exploited
syntactical relations to identify sentimental adjectives [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Turney and Littman
apply two di erent association measurements to identify new sentimental terms
in [17]. In [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] Pang and Lee present a ne-grained approach to detect the
exact sentiment (i.e. the star rating) of reviews using Support Vector Machines.
Subrahmanian and Reforgiato base sentiment detection on a syntactical level by
using adjective-verb-adjective combinations [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ].
      </p>
      <p>
        Some works also use context information to re ne sentiment indicators.
According to Nasukawa and Yi [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] sentiment detection is a three step process,
where the identi cation of sentiment expressions is followed by the
determination of their polarity and strength. The last step of the procedure identi es the
subject the sentiment terms are related to. They model such relationships for
verbs, which either directly transfer their own sentiment or another term's
sentiment to the subject. With this model they are capable of treating expressions
such as ti prevents trouble [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. The verb prevents passes the opposite
sentiment of the term trouble to the target ti. Sentence particles di erent from verbs
directly transfer their sentiment to the subject. Kim and Hove [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] specify
subjects with a Named-Entity-Recognition and assign them the overall sentiment
value of the sentence. A list of 44 verbs and 34 adjectives expanded by WordNet
[
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] synonyms and antonyms serves as sentiment lexicon. To handle complex
sentence structures such as \the California Supreme Court disagreed that the state's
new term-limit law was unconstitutional " [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] they developed a strategy, where
several negative sentiment terms in one and the same sentence eliminate each
other. Polanyi and Zaenen present a number of \contextual valence shifters" in
their eponymous work [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. Agarwal et al. propose syntactical capturing of
context in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Wilson et al. evaluate a large number of textual features, including
context, in [21] on di erent machine learning algorithms; they use a two-stage
process, rstly ltering neutral expressions from polar ones and afterwards
disambiguating the sentiment of the polar expressions. In [22] they present a similar
procedure with an expanded set of machine learners.
      </p>
      <p>
        Turney and Littman [17] use Pointwise Mutual Information (PMI) and
Latent Semantic Analysis (LSA) to identify sentiment terms in a large Web corpus.
Terms with su cient co-occurrence frequency with one of 14 paradigm terms (i.e.
a gold standard list of seven positive and negative terms) are assigned the same
sentiment value as the respective paradigm term. Evaluated on the General
Inquirer [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] PMI shows results comparable with the algorithm of Hatzivassiloglou
and McKeown [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Using three di erent extraction corpora and the sentiment
lexicon of [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] Turney and Littman show that PMI does not outperform
Hatzivassiloglou's and McKeown's algorithm but is more scalable [19]. LSA also provided
better results, but was not as scalable as PMI too. In [18] Turney uses the same
techniques to identify new sentiment terms from a paradigm list of only two
terms (excellent and poor ). This procedure performed well on the review
corpus. Beineke et al. re-interpret the previously discussed mutual association as a
Nave Bayes approach [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]; they also expand this perspective (which is an
unsupervised approach) and create a supervised approach using labeled data.
      </p>
      <p>
        Lau et al. [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] prove the importance of context by applying three di
erent language models, whereof one is an inferential language model sensible for
context. According to their evaluation the inferential language model
outperforms the other two models, emphasizing the importance of context. Bikel and
Sorensen apply a simple feature selection together with a perceptron
classier to reviews from Amazon.com [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. They use all tokens with an occurrence
frequency higher than four and achieve an accuracy of 89% in their
experiments. Denecke [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] applies a machine learning approach to multi-lingual
sentiment detection using movie reviews from six di erent languages. Google
Translator (www.google.com/language tools) translates foreign-language documents
into English. The feature selection procedure extracts a total of 77 features out
of four superclasses [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]: (1) the frequency of word classes (i.e. the number of
verbs, nouns, etc.), (2) polarity scores for the 20 most frequent words and the
averages scores for all verbs, nouns and adjectives are calculated using
SentiWordNet [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]; other features are (3) the frequency of positive and negative words
according to the General Inquirer and (4) textual features such as the number
of question marks. Using all features the Simple Logistic classi er of the WEKA
tool[
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] reaches exorbitantly good results when applied to native English
documents. When applied to non-native, translated documents the results are still
higher than the baseline demonstrating the e cacy of using a lexical resource
such as SentiWordNet.
      </p>
      <p>
        Our contextualization method is di erent from the presented context-aware
approaches. For example, we do not use linguistic relations such as synonymy
as Esuli and Sebastiani in [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Furthermore, we also do not transfer sentiment
from sentiment terms to subjects as done in [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], nor do we lter polar from
neutral expressions as or use prede ned syntactical features [21, 22]. Instead, the
proposed method considers the term's context based on discriminators identi ed
in the text and adjusts its sentiment value accordingly.
3
      </p>
    </sec>
    <sec id="sec-3">
      <title>Methodology</title>
      <p>
        The work is based on [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] and can be roughly divided into three steps (also see
Figure 3). The rst step comprises the enrichment of an initial sentiment lexicon
with contextual information. The initial lexicon is a lexicon based on sentimental
terms from the General Inquirer [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. We applied \reverse lemmatization" on
these terms, which adds in ected forms to the initial terms. The second step
is the application of the created contextualized sentiment lexicon on unknown
documents, using the Nave Bayes technique to recalculate the original sentiment
values in the sentiment lexicon. The last step comprises the identi cation of
context features applicable across the domains of the training corpora. This
step results in the creation of a generic contextualized lexicon. We compare
the improvement achieved with this approach using a lexical algorithm as our
baseline. This algorithm sums up the sentiment values of all sentiment terms
occurring in a document:
      </p>
      <p>Sent(ti) =</p>
      <p>Sent(doc) =
n
X Sent(ti)
i=1
8&gt;1; if ti is a positive term
&lt;</p>
      <p>1; if ti is a negative term
&gt;:0; if the term is neutral
by</p>
      <p>In case of a negation trigger preceding a sentiment term its value is multiplied
1. In the following, we describe each of these steps in more detail:
Generation of the contextualized lexicon The system identi es ambiguous
terms in the initial sentiment lexicon by analyzing their usage in a labeled
training set. The training set consists of documents with positive and
negative labels. A sentiment term with equally high frequency in both parts is
considered to be an ambiguous term. All ambiguous terms identi ed with
that process undergo a so-called \contextualization". This means, that the
system identi es terms frequently co-occurring with the ambiguous term in
positive/negative reviews (i.e. context terms). The contextualization creates
a contextualized lexicon. This lexicon stores the probability that a certain
ambiguous term in combination with certain context terms is normally used
in positive/negative reviews.</p>
      <p>Training
Corpus
Sentiment
Lexicon</p>
      <sec id="sec-3-1">
        <title>Generation of the Contextualized Lexicon</title>
        <p>Identifying Ambiguous Sentiment Terms
Collecting Context Information
Ambiguous</p>
        <p>Terms
Determination
Ambiguous</p>
        <p>Terms
Context Terms
Determination</p>
        <p>Contextualized</p>
        <p>Sentiment
Lexicon
Ambiguous</p>
        <p>Terms
Training
Corpus</p>
        <p>Sentiment
Value</p>
      </sec>
      <sec id="sec-3-2">
        <title>Classifying an Unknown Document</title>
        <p>Contextualized</p>
        <p>Sentiment</p>
        <p>Lexicon
Sentiment
Lexicon</p>
        <p>Naïve Bayes
Technique</p>
        <p>Test
Document
= Corpora
= Processes
= Output data
Application on unknown documents Each time a sentiment term occurs in
a new document, the contextualized sentiment lexicon is consulted and
decides, if the term is ambiguous. For non-ambiguous terms the lexicon returns
the original sentiment value of the term. In case of an ambiguous term the
system analyzes the context of the document. It uses the ten strongest
context sentiment terms and calculates the probability of the ambiguous term
being positive/negative given these ten context terms.</p>
        <p>The system calculates an ambiguous term's sentiment given context c using
the Nave Bayes formula (ci is a single context term):
p(Sent+jc) =
p(Sent+) Qin=1 p(cijSent+)</p>
        <p>Qin=1 p(ci)
The resulting value is the nal sentiment value of the ambiguous term.
Finally, the sentiment values of all sentiment terms (ambiguous and
nonambiguous) are summed up. The sum is the overall sentiment of the
document.</p>
        <p>Figure 4 shows an example of the context-sensitive sentiment detection. The
system analyzes the document and nds the sentiment term \repair", which
turns out to be ambiguous. So, it also analyzes the context, i.e. all other terms
of the document. It identi es the three context terms \friendly", \quickly",
and \reliable" as indicators for a positive meaning of \repair". Thus, the
system assigns it a positive sentiment value and classi es the whole document
as being positive. Note that the example is very simple - in reality a document
usually contains more sentiment terms, both ambiguous and non-ambiguous.</p>
        <p>Context Terms of „Repair“
Indicators for positive</p>
        <p>context
reliable long-lasting affordable
pick-up-service fast
replacementcar cooperative friendly
straightforward quickly</p>
        <p>Indicators for negative</p>
        <p>context
slowly re-do unreliable complaint</p>
        <p>slow expensive cheater wait
mistake damage replace waiting</p>
        <p>Repair
Unknown Document</p>
        <p>The service staff was
friendly. They accomplished
the repair of my car’s motor
very quickly. After driving it
for another three months I
can say that the motor is as
reliable as it was before.</p>
        <p>Context analysis using
the contextualized
lexicon</p>
        <p>Assessment:
Positive Document
Identifying Generic Features Generic features are context terms which can
be used across domains. Having obtained the contextualized lexicons from
several training corpora the system distinguishes between three types of
context term categories:
{ Helpful: Using a helpful sentiment term improves the e cacy of
sentiment detection.
{ Neutral: These terms do not change the e cacy.</p>
        <p>{ Harmful: Harmful terms reduce the e cacy.</p>
        <p>The categorization into helpful, neutral and harmful is accomplished as
follows: if a review has been classi ed incorrectly by our baseline (i.e. the lexical
algorithm explained at the beginning of this section), but correctly by the
Nave Bayes approach, the context terms of all ambiguous terms in this
document are considered as helpful terms. If it has been correctly classi ed by
the baseline but is incorrectly classi ed by Nave Bayes all context terms
are considered as harmful. Neutral context terms are those occurring in
documents where Nave Bayes and the baseline deliver the same classi cation.
Using such a procedure means that a term helpful in document A can be
neutral or even harmful in document B. A special exclusion strategy
decides which of the harmful terms should be discared, and thus also their
occurrences as helpful or neutral terms.
HelpfulA</p>
        <p>NeutralA</p>
        <p>HelpfulB</p>
        <p>NeutralB
HarmfulA</p>
        <p>HarmfulB
2 Merging
HelpfulA,
NeutralA</p>
        <p>HelpfulB,
NeutralB</p>
        <p>HarmfulA</p>
        <p>HarmfulB
3 Excluding harmful terms</p>
        <p>HarmfulA</p>
        <p>HarmfulB
HelpfulA,
NeutralA</p>
        <p>HelpfulB,</p>
        <p>
          NeutralB
We evaluated the contextualization re nements on the same corpora as in [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ],
which are a set of 2 500 products reviews from Amazon1 and 1 800 holiday reviews
from TripAdvisor2 (which we call the \Amazon" and the \TripAdvisor" corpus
later on). We accomplished a 10-fold cross-validation on both evaluation sets.
A simple lexical approach serves as the baseline for the evaluation, summing
up sentiment values of the sentiment terms occurring in the document to be
classi ed. The sentiment values come from the initial lexicon described in Section
3.
        </p>
        <p>We tested the following strategies for the exclusion of harmful terms:
{ Call: no harmful terms are excluded.
{ C n H: even terms with a single harmful occurrence are excluded.
{ C = fcj FF((ccj:jhh)) &gt; 5g: if a term has been helpful/neutral, but also has a
harmful occurrence, its frequency in helpful/neutral cases must be ve times
higher than in harmful cases.
{ C = fcj FF((ccj:jhh)) &gt; 10g: if a term has been helpful/neutral, but also has a
harmful occurrence, its frequency in helpful/neutral cases must be ten times
higher than in harmful cases.
1 amazon.com
2 tripadvisor.com
{ H: only terms with harmful occurrences are used.</p>
        <p>In Table 2 we give the results (i.e. the F-measures) for all tested exclusion
strategies. For each corpus we distinguish between positive and negative and list
the F-measure for each type (indicated by and ). The evaluation shows that
excluding harmful terms requires great care. Removing all terms with harmful
occurrences (C n H) gives worse results than leaving them untouched (Call).
Setting the ratio of non-harmful terms to harmful terms to high (i.e. &gt; 10) gives
the same results as keeping all harmful terms. Using only terms having harmful
occurrences lowers the evaluation results strongly. Yet, the results are not low
enough to judge them as completely useless. Finally, using a weaker ratio (i.e.
&gt; 5) delivers the best results.</p>
        <p>Call C n H C = fcj FF((ccj:jhh)) &gt; 5g C = fcj FF((ccj:jhh g &gt; 10 H
Amazon
The evaluation showed that particular aggregation strategies improve the overall
result for sentiment detection using contextualized lexicons. Their sole impact is
not too large, but they should be regarded as an integral component of a battery
of re nement strategies for generically contextualized sentiment detection.</p>
        <p>Future work comprises the investigation on further, more potential
aggregation strategies. Moreover, an investigation of the semantic and syntactical
sentence structure will be accomplished. The idea is that certain sentence types
might mislead sentiment detection. For example, sentences which are too short
or too long, or are in another way distorted might be counterproductive for
sentiment detection. If used anyways those sentences worsen classi cation
results. Sentiment detection would bene t from a-priori ltering of these.
Machinelearning methods can accomplish this task.
17. P.D. Turney and M.L. Littman. Unsupervised learning of semantic orientation
from a hundred-billion-word corpus. Technical report, National Research Council,</p>
        <sec id="sec-3-2-1">
          <title>Institute for Information Technology, 2002. 18. Peter D. Turney. Thumbs up or thumbs down? Semantic orientation applied to unsupervised classi cation of reviews. In ACL '02: Proceedings of the 40th Annual</title>
          <p>Meeting on Association for Computational Linguistics, pages 417{424, Morristown,</p>
        </sec>
        <sec id="sec-3-2-2">
          <title>NJ, USA, 2002. Association for Computational Linguistics.</title>
          <p>19. Peter D. Turney and Michael L. Littman. Measuring praise and criticism:
Inference of semantic orientation from association. ACM Transactions on Information
Systems, 21(4):315{346, 2003.
20. Janyce M. Wiebe. Tracking point of view in narrative. Computational Linguistics,
20(2):233{287, 1994.
21. Theresa Wilson, Janyce Wiebe, and Paul Ho mann. Recognizing contextual
polarity in phrase-level sentiment analysis. In HLT '05: Proceedings of the conference on
Human Language Technology and Empirical Methods in Natural Language
Processing, pages 347{354, Morristown, NJ, USA, 2005. Association for Computational</p>
        </sec>
        <sec id="sec-3-2-3">
          <title>Linguistics.</title>
          <p>22. Theresa Wilson, Janyce Wiebe, and Paul Ho mann. Recognizing contextual
polarity: An exploration of features for phrase-level sentiment analysis. Computational
Linguistics, 35(3):399{433, 2009.</p>
        </sec>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <given-names>Apoorv</given-names>
            <surname>Agarwal</surname>
          </string-name>
          , Fadi Biadsy, and
          <string-name>
            <surname>Kathleen</surname>
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Mckeown</surname>
          </string-name>
          .
          <article-title>Contextual phrase-level polarity analysis using lexical a ect scoring and syntactic N-grams</article-title>
          .
          <source>In EACL '09: Proceedings of the 12th Conference of the European Chapter of the Association for Computational Linguistics</source>
          , pages
          <volume>24</volume>
          {
          <fpage>32</fpage>
          ,
          <string-name>
            <surname>Morristown</surname>
          </string-name>
          , NJ, USA,
          <year>2009</year>
          .
          <article-title>Association for Computational Linguistics</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <given-names>Philip</given-names>
            <surname>Beineke</surname>
          </string-name>
          , Trevor Hastie, and
          <string-name>
            <given-names>Shivakumar</given-names>
            <surname>Vaithyanathan</surname>
          </string-name>
          .
          <article-title>The sentimental factor: Improving review classi cation via human-provided information</article-title>
          .
          <source>In ACL '04: Proceedings of the 42nd Annual Meeting on Association for Computational Linguistics</source>
          , page
          <volume>263</volume>
          ,
          <string-name>
            <surname>Morristown</surname>
          </string-name>
          , NJ, USA,
          <year>2004</year>
          .
          <article-title>Association for Computational Linguistics</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Daniel</surname>
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Bikel</surname>
          </string-name>
          and
          <article-title>Je rey Sorensen. If we want your opinion</article-title>
          .
          <source>In ICSC 2007. International Conference on Semantic Computing</source>
          , pages
          <volume>493</volume>
          {
          <fpage>500</fpage>
          , Irvine, CA,
          <year>September 2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <given-names>Kerstin</given-names>
            <surname>Denecke</surname>
          </string-name>
          .
          <article-title>How to assess customer opinions beyond language barriers?</article-title>
          <source>In Third International Conference on Digital Information Management</source>
          , pages
          <volume>430</volume>
          {
          <fpage>435</fpage>
          . IEEE,
          <year>November 2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <given-names>Andrea</given-names>
            <surname>Esuli</surname>
          </string-name>
          and
          <string-name>
            <given-names>Fabrizio</given-names>
            <surname>Sebastiani</surname>
          </string-name>
          .
          <article-title>SentiWordNet: A publicly available lexical resource for opinion mining</article-title>
          .
          <source>In Proceedings of the 5th International Conference on Language Resources and Evaluation</source>
          , pages
          <volume>417</volume>
          {
          <fpage>422</fpage>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <given-names>C.</given-names>
            <surname>Fellbaum. WordNet -</surname>
          </string-name>
          <article-title>An electronic lexical database</article-title>
          .
          <source>Computational Linguistics</source>
          ,
          <volume>25</volume>
          (
          <issue>2</issue>
          ):
          <volume>292</volume>
          {
          <fpage>296</fpage>
          ,
          <year>1998</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Stefan</surname>
            <given-names>Gindl</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Albert</given-names>
            <surname>Weichselbraun</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Arno</given-names>
            <surname>Scharl</surname>
          </string-name>
          .
          <article-title>Cross-domain contextualization of sentiment lexicons</article-title>
          .
          <source>In ECAI 2010: Proceedings of the 19th European Conference on Arti cial Intelligence</source>
          , in press.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <given-names>Mark</given-names>
            <surname>Hall</surname>
          </string-name>
          , Eibe Frank, Geo rey Holmes, Bernhard Pfahringer,
          <string-name>
            <given-names>Peter</given-names>
            <surname>Reutemann</surname>
          </string-name>
          , and
          <string-name>
            <surname>Ian</surname>
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Witten</surname>
          </string-name>
          .
          <article-title>The WEKA data mining software: An update</article-title>
          .
          <source>SIGKDD Explorations</source>
          ,
          <volume>11</volume>
          (
          <issue>1</issue>
          ):
          <volume>10</volume>
          {
          <fpage>18</fpage>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <given-names>Vasileios</given-names>
            <surname>Hatzivassiloglou</surname>
          </string-name>
          and
          <string-name>
            <surname>Kathleen R. McKeown</surname>
          </string-name>
          .
          <article-title>Predicting the semantic orientation of adjectives</article-title>
          .
          <source>In Proceedings of the eighth conference on European chapter of the Association for Computational Linguistics</source>
          , pages
          <volume>174</volume>
          {
          <fpage>181</fpage>
          ,
          <string-name>
            <surname>Morristown</surname>
          </string-name>
          , NJ, USA,
          <year>1997</year>
          .
          <article-title>Association for Computational Linguistics</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Soo-Min Kim</surname>
            and
            <given-names>Eduard</given-names>
          </string-name>
          <string-name>
            <surname>Hovy</surname>
          </string-name>
          .
          <article-title>Determining the sentiment of opinions</article-title>
          .
          <source>In COLING '04: Proceedings of the 20th international conference on Computational Linguistics</source>
          , page
          <volume>1367</volume>
          ,
          <string-name>
            <surname>Morristown</surname>
          </string-name>
          , NJ, USA,
          <year>2004</year>
          .
          <article-title>Association for Computational Linguistics</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>R.Y.K. Lau</surname>
            ,
            <given-names>C.L.</given-names>
          </string-name>
          <string-name>
            <surname>Lai</surname>
            , and
            <given-names>Yuefeng</given-names>
          </string-name>
          <string-name>
            <surname>Li</surname>
          </string-name>
          .
          <article-title>Leveraging the web context for contextsensitive opinion mining</article-title>
          .
          <source>In Computer Science and Information Technology</source>
          ,
          <year>2009</year>
          .
          <source>ICCSIT</source>
          <year>2009</year>
          . 2nd IEEE International Conference on, pages
          <volume>467</volume>
          {
          <fpage>471</fpage>
          ,
          <string-name>
            <surname>Aug</surname>
          </string-name>
          .
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <given-names>Tetsuya</given-names>
            <surname>Nasukawa</surname>
          </string-name>
          and
          <string-name>
            <given-names>Jeonghee</given-names>
            <surname>Yi</surname>
          </string-name>
          .
          <article-title>Sentiment analysis: Capturing favorability using natural language processing</article-title>
          .
          <source>In K-CAP '03: Proceedings of the 2nd international conference on Knowledge capture</source>
          , pages
          <volume>70</volume>
          {
          <fpage>77</fpage>
          , New York, NY, USA,
          <year>2003</year>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <given-names>Bo</given-names>
            <surname>Pang</surname>
          </string-name>
          and
          <string-name>
            <given-names>Lillian</given-names>
            <surname>Lee</surname>
          </string-name>
          .
          <article-title>Seeing stars: Exploiting class relationships for sentiment categorization with respect to rating scales</article-title>
          .
          <source>In ACL '05: Proceedings of the 43rd Annual Meeting on Association for Computational Linguistics</source>
          , pages
          <volume>115</volume>
          {
          <fpage>124</fpage>
          ,
          <string-name>
            <surname>Morristown</surname>
          </string-name>
          , NJ, USA,
          <year>2005</year>
          .
          <article-title>Association for Computational Linguistics</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <given-names>Livia</given-names>
            <surname>Polanyi</surname>
          </string-name>
          and
          <string-name>
            <given-names>Annie</given-names>
            <surname>Zaenen</surname>
          </string-name>
          .
          <article-title>Contextual valence shifters</article-title>
          .
          <source>In Computing Attitude and A ect in Text: Theory and Applications</source>
          ,
          <source>The Information Retrieval Series</source>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Philip J. Stone</surname>
          </string-name>
          ,
          <string-name>
            <surname>Dexter C. Dunphy</surname>
          </string-name>
          , and
          <string-name>
            <surname>Marshall</surname>
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Smith. The General</surname>
          </string-name>
          <article-title>Inquirer: A computer approach to content analysis. M.I.T</article-title>
          . Press, Cambridge, Massachusetts,
          <year>1966</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16. V.S. Subrahmanian and Diego Reforgiato. AVA:
          <article-title>Adjective-Verb-Adverb combinations for sentiment analysis</article-title>
          .
          <source>Intelligent Systems</source>
          , IEEE,
          <volume>23</volume>
          (
          <issue>4</issue>
          ):
          <volume>43</volume>
          {
          <fpage>50</fpage>
          ,
          <string-name>
            <surname>July</surname>
            <given-names>-August</given-names>
          </string-name>
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>