<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Concept Term Expansion Approach for Monitoring Reputation of Companies on Twitter</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>M. Atif Qureshi</string-name>
          <email>muhammad.qureshi@nuigalway.ie</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Colm O'Riordan</string-name>
          <email>colm.oriordan@nuigalway.ie</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Gabriella Pasi</string-name>
          <email>pasi@disco.unimib.it</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Computational Intelligence Research Group, National University of Ireland Galway</institution>
          ,
          <country country="IE">Ireland</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Information Retrieval Lab</institution>
          ,
          <addr-line>Informatics, Systems and Communication</addr-line>
          ,
          <institution>University of Milan Bicocca</institution>
          ,
          <addr-line>Milan</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2012</year>
      </pub-date>
      <abstract>
        <p>The aim of this contribution is to easily monitor the reputation of a company in the Twittersphere. We propose a strategy that organizes a stream of tweets into different clusters based on the tweets topics. Furthermore, the obtained clusters are prioritized into different priority levels. A cluster with high priority represents a topic which may affect the reputation of a company, and that consequently deserves immediate attention. The evaluation results show that our method is competitive even though the method does not make use of any external knowledge resource.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Twitter3 has become an immensely popular microblogging platform with over
140M unique visitors, and around 340M tweets per day4. Based on its
growing popularity, several companies have started to use Twitter as a medium for
electronic word-of-mouth marketing [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. There is also an increasing trend of
Twitter users to express their opinions about various companies and their
products via tweets. Hence, tweets serve as a significant repository for a company to
monitor its online reputation; this motivates the need to take the necessary steps
to tackle threats to it. This, however, involves considerable research challenges
and motivates the research reported in our paper, the main characteristics of
which are:
– clustering tweets based on their topics: for example, the company Apple
would have separate topical clusters for iPhone, iPad and iPod etc.
– ordering tweets by priority to the company: the idea is that tweets critical
to the company’s reputation require an immediate action, and they have a
higher priority than tweets that do not require immediate attention. For
example, a tweet heavily criticizing a company’s customer service may damage
the company’s reputation, and thus they should have high priority.
      </p>
      <sec id="sec-1-1">
        <title>3 http://twitter.com</title>
        <p>4 http://blog.twitter.com/2012/03/twitter-turns-six.html</p>
        <p>In this paper, we focus on the task of monitoring tweets for a company’s
reputation, in the context of the RepLab2012, where we are given a set of companies
and for each company a set of tweets, which contain different topics pertaining
to the company with different levels of priority. Performing such a monitoring of
tweets is a significantly challenging task as tweet messages are very short (140
characters) and noisy. We alleviate these problems through the idea of concept
term expansion in tweets. We perform two phases of clustering and priority level
assessment separately, whereby the clustering employs unsupervised techniques
while supervision is used for priority level assessment.</p>
        <p>The rest of the paper is organized as follows. Section 2 describes the problem
in more detail. Section 3 presents our technique for clustering and assigning
priority levels to the clusters. Section 4 describes the experiments and finally
Section 5 concludes the paper.
2</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>Problem</title>
    </sec>
    <sec id="sec-3">
      <title>Description</title>
      <p>In this section, we briefly define the problem statement related to this
contribution. We were provided with a stream of tweets for different companies collected
by issuing a query corresponding to the company name. The stream of tweets
for the companies were then divided into a training set and a test set. In the
training set each stream of tweets for a company was clustered according to their
topics. Furthermore, these clusters were prioritized into five different levels as
follows:</p>
      <p>Alert &gt;average priority &gt;low priority &gt;‘other’ cluster &gt;‘irrelevant’
Alert corresponds to the cluster of tweets that deserves immediate
attention by the company. Likewise, the tweet clusters with average and low priority
deserve attention as well but relatively less attention than those with alert
priority level. In the case of ‘other’, these are clusters of tweets that are about the
company but that do not qualify as interesting topics and are negligible to the
monitoring purposes. Finally, ‘irrelevant’ are the cluster of tweets that do not
refer to the company.</p>
      <p>Our task is to cluster the stream of unseen tweets (test set) of a given
company with respect to topics. Furthermore, we have to assign these clusters a
priority level chosen from the above-mentioned five levels.
3</p>
    </sec>
    <sec id="sec-4">
      <title>Methodology</title>
      <p>The proposed method is entirely based on the tweets contents, i.e., it does not use
any external knowledge resource such as Wikipedia or the content of any Web
page. Before applying our method we expand the shortened URL mentioned
inside a tweet into a full URL in order to avoid redundant link forwarders to
the same URL. Furthermore, a tweet that is not written in English is translated
into the English Language by using the Bing Translation API5. In the following
subsections we present the proposed strategy to analyse tweets.</p>
      <sec id="sec-4-1">
        <title>5 http://www.microsofttranslator.com/</title>
        <p>3.1</p>
        <sec id="sec-4-1-1">
          <title>Tweet concept terms extraction</title>
          <p>
            In the first step, we extract concept terms (i.e., important terms) from each
tweet so as to be able to identify a topic. To achieve this goal, we filter out the
trivial components from each tweet’s content such as mentions, RT, MT and
URL string. Then, we apply POS tagging [
            <xref ref-type="bibr" rid="ref5">5</xref>
            ] to identify the terms having a label
‘NN’, ‘NNS’, ‘NNP’, ‘NNPS’, ‘JJ’ or ‘CD’ as a concept term.
3.2
          </p>
        </sec>
        <sec id="sec-4-1-2">
          <title>Training priority scores for concept terms</title>
          <p>In this step, to a concept term multiple weights are assigned, which describe the
strength of association of a concept term with each priority level. To this aim,
we employ the training data in which each cluster of tweets is labelled with a
priority level. Each tweet in a cluster is associated with the label of that cluster
i.e., we borrow the label from the tweet’s cluster. After this, we assign a score to
each concept term corresponding to its strength of association with each priority
level (borrowed from the tweet’s label). For example, a concept term mentioned
frequently in tweets that have a specific priority level gets the highest score
for that particular priority level, while the same concept term when mentioned
rarely in other tweets labelled with a different priority level would get a low score
for this priority level.
3.3</p>
        </sec>
        <sec id="sec-4-1-3">
          <title>Main algorithm</title>
          <p>In this section we describe the main algorithm that clusters the stream of tweets
with respect to their topics, and which assigns them a priority level. The
algorithm iteratively learns two threshold values (i.e., the content threshold and the
specificity threshold) from a list of threshold values provided to it as explained
in the following sections.
3.3.1 Clustering In this step, we cluster tweets according to their contents
similarity, to the specificity of concept terms used among tweets, and to
common URL mentions among the tweets. For detecting contents similarity we used
the content threshold, and for determining the specificity of concept terms we
used the specificity threshold. After this step we have all the tweets clustered
according to their main topics.
3.3.2 Predicting priority levels In this step, we assign a priority level to
each cluster. To this aim, we first estimate a priority level for each tweet in the
corpus, and then by using the assignment of priority level to the tweets we decide
a priority for each cluster. The process is explained here below.</p>
          <p>Estimate of priority level for each tweet</p>
          <p>First, we generate five aggregations across each priority level for a tweet.
Then, the highest aggregation corresponding to a priority level becomes the
priority level for that tweet. Each aggregation is computed by aggregating each
concept term’s priority score (as estimated in Section 3.2) corresponding to the
priority level for that tweet.</p>
          <p>Estimate of priority level for each cluster</p>
          <p>Since each cluster is composed of tweets, the assigned priority level for a
tweet is counted as a vote for a cluster’s priority level, and the priority level that
gets the maximum number of votes (for a cluster) becomes the priority level for
that cluster.
3.3.3 Global error estimate and optimization This step enables the
algorithm to learn optimized threshold values. To this aim, we estimate the global
error as follows. We first estimate the number of errors per cluster by counting
the number of inconsistencies (i.e., non-uniformity) among the priority levels
assigned to the tweets of a cluster. Then, we aggregate these errors estimates across
each cluster to define a global error estimate. The threshold values across which
the global error estimation is minimum are declared to be optimized threshold
values. The output corresponding to the optimized threshold values is reported
as the final output of the algorithm.
4
4.1</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Experimental Results</title>
      <sec id="sec-5-1">
        <title>Data set</title>
        <p>
          We performed our experiments by using the data set provided by the Monitoring
task of RepLab 2012 [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]. In this data set 37 companies were provided, six out
of which in the training set, while the remaining 31 in the test set. For each
company a few hundred tweets were provided.
4.2
        </p>
      </sec>
      <sec id="sec-5-2">
        <title>Evaluation Measures</title>
        <p>
          The measures used to the evaluation purposes are Reliability and Sensitivity,
which are described in detail in [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ].
        </p>
        <p>
          In essence, these measures consider two types of binary relationships
between pairs of items: relatedness – two items belong to the same cluster – and
priority – one item has more priority than the other. Reliability is defined as
precision of binary relationships predicted by the system with respect to those
that derive from the gold standard. Sensitivity is similarly defined as the recall
of relationships. When only clustering relationships are considered, Reliability
and Sensitivity are equivalent to BCubed Precision and Recall [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ].
4.3
        </p>
      </sec>
      <sec id="sec-5-3">
        <title>Results</title>
        <p>Table 1 presents a snapshot of the official results for the Monitoring task of
RepLab 2012, where CIRGDISCO is the name of our team.</p>
        <p>Table 1 shows that our algorithm performed competitively, and it is the
second from the top; it is important to notice that our algorithm did not use
R</p>
        <p>S</p>
        <p>F(R,S)
R Clustering S Clustering F(R,S) R S F
(BCubed (BCubed Clustering Prior Priority Priority
precision) recall)
0.32
0.24
0.34
0.34
0.2
any external knowledge resource although sources of evidence were provided in
the data set. The main reason for not using these resources was the shortage of
time; this means that there is a natural room for improvement of our algorithm
and for further investigation. In addition, our algorithm shows the best BCubed
precision compared to other algorithms.
5</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Conclusion</title>
      <p>We proposed an algorithm for clustering tweets and for assigning them a priority
level for companies. Our algorithm did not make use of any external knowledge
resource and did not require prior information about the company. Even under
these constraints our algorithm showed competitive performance. However, there
is a room for improvements where external evidence could play a promising added
value.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1. E. Amigo´,
          <string-name>
            <given-names>A.</given-names>
            <surname>Corujo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gonzalo</surname>
          </string-name>
          , E. Meij, and M. d. Rijke. Overview of replab 2012:
          <article-title>Evaluating online reputation management systems</article-title>
          .
          <source>In CLEF 2012 Labs and Workshop Notebook Papers</source>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <given-names>E.</given-names>
            <surname>Amigo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gonzalo</surname>
          </string-name>
          , and
          <string-name>
            <given-names>F.</given-names>
            <surname>Verdejo</surname>
          </string-name>
          .
          <article-title>Reliability and Sensitivity: Generic Evaluation Measures for Document Organization Tasks</article-title>
          . UNED, Madrid, Spain,
          <year>2012</year>
          .
          <source>Technical Report.</source>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <given-names>B. J.</given-names>
            <surname>Jansen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Sobel</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Chowdury</surname>
          </string-name>
          .
          <article-title>Micro-blogging as online word of mouth branding</article-title>
          .
          <source>In Proceedings of the 27th international conference extended abstracts on Human factors in computing systems, CHI EA '09</source>
          , pages
          <fpage>3859</fpage>
          -
          <lpage>3864</lpage>
          , New York, NY, USA,
          <year>2009</year>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <given-names>B. J.</given-names>
            <surname>Jansen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Sobel</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Chowdury</surname>
          </string-name>
          .
          <article-title>Twitter power: Tweets as electronic word of mouth</article-title>
          .
          <source>J. Am. Soc. Inf. Sci. Technol</source>
          .,
          <volume>60</volume>
          (
          <issue>11</issue>
          ):
          <fpage>2169</fpage>
          -
          <lpage>2188</lpage>
          , Nov.
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Klein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. D.</given-names>
            <surname>Manning</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Singer</surname>
          </string-name>
          .
          <article-title>Feature-rich part-of-speech tagging with a cyclic dependency network</article-title>
          .
          <source>In Proceedings of the 2003 Conference of the North American Chapter of the Association for Computational Linguistics on Human Language Technology - Volume 1, NAACL '03</source>
          , pages
          <fpage>173</fpage>
          -
          <lpage>180</lpage>
          , Stroudsburg, PA, USA,
          <year>2003</year>
          .
          <article-title>Association for Computational Linguistics</article-title>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>