=Paper=
{{Paper
|id=Vol-1178/CLEF2012wn-RepLab-QureshiEt2012
|storemode=property
|title=Concept Term Expansion Approach for Monitoring Reputation of Companies on Twitter
|pdfUrl=https://ceur-ws.org/Vol-1178/CLEF2012wn-RepLab-QureshiEt2012.pdf
|volume=Vol-1178
}}
==Concept Term Expansion Approach for Monitoring Reputation of Companies on Twitter==
Concept Term Expansion Approach for Monitoring Reputation of Companies on Twitter M. Atif Qureshi1,2 , Colm O’Riordan1, and Gabriella Pasi2 1 Computational Intelligence Research Group, National University of Ireland Galway, Ireland 2 Information Retrieval Lab,Informatics, Systems and Communication, University of Milan Bicocca, Milan, Italy muhammad.qureshi@nuigalway.ie, colm.oriordan@nuigalway.ie, pasi@disco.unimib.it Abstract. The aim of this contribution is to easily monitor the reputa- tion of a company in the Twittersphere. We propose a strategy that orga- nizes a stream of tweets into different clusters based on the tweets topics. Furthermore, the obtained clusters are prioritized into different priority levels. A cluster with high priority represents a topic which may affect the reputation of a company, and that consequently deserves immediate attention. The evaluation results show that our method is competitive even though the method does not make use of any external knowledge resource. 1 Introduction Twitter3 has become an immensely popular microblogging platform with over 140M unique visitors, and around 340M tweets per day4 . Based on its grow- ing popularity, several companies have started to use Twitter as a medium for electronic word-of-mouth marketing [3] [4]. There is also an increasing trend of Twitter users to express their opinions about various companies and their prod- ucts via tweets. Hence, tweets serve as a significant repository for a company to monitor its online reputation; this motivates the need to take the necessary steps to tackle threats to it. This, however, involves considerable research challenges and motivates the research reported in our paper, the main characteristics of which are: – clustering tweets based on their topics: for example, the company Apple would have separate topical clusters for iPhone, iPad and iPod etc. – ordering tweets by priority to the company: the idea is that tweets critical to the company’s reputation require an immediate action, and they have a higher priority than tweets that do not require immediate attention. For ex- ample, a tweet heavily criticizing a company’s customer service may damage the company’s reputation, and thus they should have high priority. 3 http://twitter.com 4 http://blog.twitter.com/2012/03/twitter-turns-six.html In this paper, we focus on the task of monitoring tweets for a company’s rep- utation, in the context of the RepLab2012, where we are given a set of companies and for each company a set of tweets, which contain different topics pertaining to the company with different levels of priority. Performing such a monitoring of tweets is a significantly challenging task as tweet messages are very short (140 characters) and noisy. We alleviate these problems through the idea of concept term expansion in tweets. We perform two phases of clustering and priority level assessment separately, whereby the clustering employs unsupervised techniques while supervision is used for priority level assessment. The rest of the paper is organized as follows. Section 2 describes the problem in more detail. Section 3 presents our technique for clustering and assigning priority levels to the clusters. Section 4 describes the experiments and finally Section 5 concludes the paper. 2 Problem Description In this section, we briefly define the problem statement related to this contribu- tion. We were provided with a stream of tweets for different companies collected by issuing a query corresponding to the company name. The stream of tweets for the companies were then divided into a training set and a test set. In the training set each stream of tweets for a company was clustered according to their topics. Furthermore, these clusters were prioritized into five different levels as follows: Alert >average priority >low priority >‘other’ cluster >‘irrelevant’ Alert corresponds to the cluster of tweets that deserves immediate atten- tion by the company. Likewise, the tweet clusters with average and low priority deserve attention as well but relatively less attention than those with alert pri- ority level. In the case of ‘other’, these are clusters of tweets that are about the company but that do not qualify as interesting topics and are negligible to the monitoring purposes. Finally, ‘irrelevant’ are the cluster of tweets that do not refer to the company. Our task is to cluster the stream of unseen tweets (test set) of a given com- pany with respect to topics. Furthermore, we have to assign these clusters a priority level chosen from the above-mentioned five levels. 3 Methodology The proposed method is entirely based on the tweets contents, i.e., it does not use any external knowledge resource such as Wikipedia or the content of any Web page. Before applying our method we expand the shortened URL mentioned inside a tweet into a full URL in order to avoid redundant link forwarders to the same URL. Furthermore, a tweet that is not written in English is translated into the English Language by using the Bing Translation API5 . In the following subsections we present the proposed strategy to analyse tweets. 5 http://www.microsofttranslator.com/ 3.1 Tweet concept terms extraction In the first step, we extract concept terms (i.e., important terms) from each tweet so as to be able to identify a topic. To achieve this goal, we filter out the trivial components from each tweet’s content such as mentions, RT, MT and URL string. Then, we apply POS tagging [5] to identify the terms having a label ‘NN’, ‘NNS’, ‘NNP’, ‘NNPS’, ‘JJ’ or ‘CD’ as a concept term. 3.2 Training priority scores for concept terms In this step, to a concept term multiple weights are assigned, which describe the strength of association of a concept term with each priority level. To this aim, we employ the training data in which each cluster of tweets is labelled with a priority level. Each tweet in a cluster is associated with the label of that cluster i.e., we borrow the label from the tweet’s cluster. After this, we assign a score to each concept term corresponding to its strength of association with each priority level (borrowed from the tweet’s label). For example, a concept term mentioned frequently in tweets that have a specific priority level gets the highest score for that particular priority level, while the same concept term when mentioned rarely in other tweets labelled with a different priority level would get a low score for this priority level. 3.3 Main algorithm In this section we describe the main algorithm that clusters the stream of tweets with respect to their topics, and which assigns them a priority level. The algo- rithm iteratively learns two threshold values (i.e., the content threshold and the specificity threshold) from a list of threshold values provided to it as explained in the following sections. 3.3.1 Clustering In this step, we cluster tweets according to their contents similarity, to the specificity of concept terms used among tweets, and to com- mon URL mentions among the tweets. For detecting contents similarity we used the content threshold, and for determining the specificity of concept terms we used the specificity threshold. After this step we have all the tweets clustered according to their main topics. 3.3.2 Predicting priority levels In this step, we assign a priority level to each cluster. To this aim, we first estimate a priority level for each tweet in the corpus, and then by using the assignment of priority level to the tweets we decide a priority for each cluster. The process is explained here below. Estimate of priority level for each tweet First, we generate five aggregations across each priority level for a tweet. Then, the highest aggregation corresponding to a priority level becomes the priority level for that tweet. Each aggregation is computed by aggregating each concept term’s priority score (as estimated in Section 3.2) corresponding to the priority level for that tweet. Estimate of priority level for each cluster Since each cluster is composed of tweets, the assigned priority level for a tweet is counted as a vote for a cluster’s priority level, and the priority level that gets the maximum number of votes (for a cluster) becomes the priority level for that cluster. 3.3.3 Global error estimate and optimization This step enables the algo- rithm to learn optimized threshold values. To this aim, we estimate the global error as follows. We first estimate the number of errors per cluster by counting the number of inconsistencies (i.e., non-uniformity) among the priority levels as- signed to the tweets of a cluster. Then, we aggregate these errors estimates across each cluster to define a global error estimate. The threshold values across which the global error estimation is minimum are declared to be optimized threshold values. The output corresponding to the optimized threshold values is reported as the final output of the algorithm. 4 Experimental Results 4.1 Data set We performed our experiments by using the data set provided by the Monitoring task of RepLab 2012 [1]. In this data set 37 companies were provided, six out of which in the training set, while the remaining 31 in the test set. For each company a few hundred tweets were provided. 4.2 Evaluation Measures The measures used to the evaluation purposes are Reliability and Sensitivity, which are described in detail in [2]. In essence, these measures consider two types of binary relationships be- tween pairs of items: relatedness – two items belong to the same cluster – and priority – one item has more priority than the other. Reliability is defined as precision of binary relationships predicted by the system with respect to those that derive from the gold standard. Sensitivity is similarly defined as the recall of relationships. When only clustering relationships are considered, Reliability and Sensitivity are equivalent to BCubed Precision and Recall [1]. 4.3 Results Table 1 presents a snapshot of the official results for the Monitoring task of RepLab 2012, where CIRGDISCO is the name of our team. Table 1 shows that our algorithm performed competitively, and it is the second from the top; it is important to notice that our algorithm did not use Table 1. Results of Monitoring task of RepLab 2012 Team R Clustering S Clustering F(R,S) R S F R S F(R,S) (BCubed (BCubed Clustering Prior Priority Priority precision) recall) UNED 3 0.72 0.32 0.4 0.25 0.3 0.26 0.32 0.26 0.29 CIRGDISCO 0.95 0.24 0.35 0.24 0.3 0.24 0.29 0.22 0.25 OPTAH 1 0.7 0.34 0.38 0.19 0.16 0.16 0.37 0.19 0.22 UNED 2 0.85 0.34 0.39 0 0 0 0.85 0.09 0.14 UNED 1 0.9 0.2 0.3 0 0 0 0.9 0.05 0.1 any external knowledge resource although sources of evidence were provided in the data set. The main reason for not using these resources was the shortage of time; this means that there is a natural room for improvement of our algorithm and for further investigation. In addition, our algorithm shows the best BCubed precision compared to other algorithms. 5 Conclusion We proposed an algorithm for clustering tweets and for assigning them a priority level for companies. Our algorithm did not make use of any external knowledge resource and did not require prior information about the company. Even under these constraints our algorithm showed competitive performance. However, there is a room for improvements where external evidence could play a promising added value. References 1. E. Amigó, A. Corujo, J. Gonzalo, E. Meij, and M. d. Rijke. Overview of replab 2012: Evaluating online reputation management systems. In CLEF 2012 Labs and Workshop Notebook Papers, 2012. 2. E. Amigo, J. Gonzalo, and F. Verdejo. Reliability and Sensitivity: Generic Eval- uation Measures for Document Organization Tasks. UNED, Madrid, Spain, 2012. Technical Report. 3. B. J. Jansen, M. Zhang, K. Sobel, and A. Chowdury. Micro-blogging as online word of mouth branding. In Proceedings of the 27th international conference extended abstracts on Human factors in computing systems, CHI EA ’09, pages 3859–3864, New York, NY, USA, 2009. ACM. 4. B. J. Jansen, M. Zhang, K. Sobel, and A. Chowdury. Twitter power: Tweets as electronic word of mouth. J. Am. Soc. Inf. Sci. Technol., 60(11):2169–2188, Nov. 2009. 5. K. Toutanova, D. Klein, C. D. Manning, and Y. Singer. Feature-rich part-of-speech tagging with a cyclic dependency network. In Proceedings of the 2003 Conference of the North American Chapter of the Association for Computational Linguistics on Human Language Technology - Volume 1, NAACL ’03, pages 173–180, Stroudsburg, PA, USA, 2003. Association for Computational Linguistics.