<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Mining Newsworthy Topics from Social Media</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Carlos Martin</string-name>
          <email>c.j.martin-dancausa@rgu.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>David Corney</string-name>
          <email>d.p.a.corney@rgu.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ayse G¨oker</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andrew MacFarlane</string-name>
          <email>a.macfarlane-1@city.ac.uk</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>IDEAS Research Institute, School of Computing &amp; Digital Media, Robert Gordon University</institution>
          ,
          <addr-line>Aberdeen AB10 7QB</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>School of Informatics, City University London</institution>
          ,
          <addr-line>London EC1V 0HB</addr-line>
        </aff>
      </contrib-group>
      <fpage>35</fpage>
      <lpage>46</lpage>
      <abstract>
        <p>Newsworthy stories are increasingly being shared through social networking platforms such as Twitter and Reddit, and journalists now use them to rapidly discover stories and eye-witness accounts. We present a technique that detects “bursts” of phrases on Twitter that is designed for a real-time topic-detection system. We describe a time-dependent variant of the classic tf-idf approach and group together bursty phrases that often appear in the same messages in order to identify emerging topics. We demonstrate our methods by analysing tweets corresponding to events drawn from the worlds of politics and sport. We created a user-centred “ground truth” to evaluate our methods, based on mainstream media accounts of the events. This helps ensure our methods remain practical. We compare several clustering and topic ranking methods to discover the characteristics of news-related collections, and show that different strategies are needed to detect emerging topics within them. We show that our methods successfully detect a range of different topics for each event and can retrieve messages (for example, tweets) that represent each topic for the user.</p>
      </abstract>
      <kwd-group>
        <kwd>topic detection</kwd>
        <kwd>Twitter</kwd>
        <kwd>temporal analysis</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>The growth of social networking sites, such as Twitter, Facebook and Reddit, is
well documented. Every day, a huge variety of information on different topics is
shared by many people. Given the real-time, global nature of these sites, they
are used by many people as a primary source of news content [1]. Increasingly,
such sites are also used by journalists, partly to find and track breaking news but
also to find user-generated content such as photos and videos, to enhance their
stories. These often come from eye-witnesses who would be otherwise difficult to
find, especially given the volume of content being shared.</p>
      <p>Our overall goal is to produce a practical tool to help journalists and news
readers to find newsworthy topics from message streams without being
overwhelmed. Note that it is not our intention to re-create Twitter’s own “trending
topics” functionality. That is usually dominated by very high-level topics and
memes, defined by just one or two words or a name and with no emphasis on
‘news’.</p>
      <p>
        Our system works by identifying phrases that show a sudden increase in
frequency (a “burst”) and then finding co-occurring groups to identify topics. Such
bursts are typically responses to real-world events. In this way, the news
consumer can avoid being overwhelmed by redundant messages, even if the initial
stream is formed of diverse messages. The emphasis is on the temporal nature
of message streams as we bring to the surface groups of messages that
contain suddenly-popular phrases. An early version of this approach was recently
described [
        <xref ref-type="bibr" rid="ref1 ref2">2, 3</xref>
        ], where it compared favourably to several alternatives and
benchmarks. Here we expand and update that work, examining the effect of different
clustering and topic ranking approaches used to form coherent topics from bursty
phrases.
2
      </p>
      <p>
        Related Work
Newman [
        <xref ref-type="bibr" rid="ref3">4</xref>
        ] discusses the central use of social media by news professionals,
such as hosting live blogs of ongoing events. He also describes the growth of
collaborative, networked journalism, where news professionals draw together a
wide range of images, videos and text from social networks and provide a curation
service. Broadcasters and newspapers can also use social media to increase brand
loyalty across a fragmented media marketplace.
      </p>
      <p>
        Petrovic et al. [
        <xref ref-type="bibr" rid="ref4">5</xref>
        ] focus on the task of first-story detection (FSD), which they
also call “new event detection”. They use a locality sensitive hashing technique
on 160 million Twitter posts, hashing incoming tweet vectors into buckets in
order to find the nearest neighbour and hence detect new events and track them.
This work is extended in Petrovic et al. [
        <xref ref-type="bibr" rid="ref5">6</xref>
        ] using paraphrases for first story
detection on 50 million tweets. Their FSD evaluation used newswire sources
rather than Tweets, based on the existing TDT5 datasets. The Twitter-based
evaluation was limited to calculating the average precision of their system, by
getting two human annotators to label the output as being about an event or
not. This contrasts with our goal here, which is to measure the topic-level recall,
i.e. to count how many newsworthy stories the system retrieved.
      </p>
      <p>
        Benhardus [
        <xref ref-type="bibr" rid="ref6">7</xref>
        ] uses standard collection statistics such as tf-idf, unigrams and
bigrams to detect trending topics. Two data collections are used, one from the
Twitter API and the second being the Edinburgh Twitter corpus containing 97
million tweets, which was used as a baseline with some natural language
processing used (e.g. detecting prepositions or conjunctions). The research focused
on general trending topics (typically finding personalities and for new hashtags)
rather than focusing the needs of journalistic users and news readers.
      </p>
      <p>
        Shamma et al. [
        <xref ref-type="bibr" rid="ref7">8</xref>
        ] focus on “peaky topics” (topics that show highly localized,
momentary interest) by using unigrams only. The focus of the method is to
obtain peak terms for a given time slot when compared to the whole corpus
rather than over a given time-frame. The use of the whole corpus favours
batchmode processing and is less suitable for real-time and user-centred analysis.
      </p>
      <p>
        Phuvipadawat and Murata [
        <xref ref-type="bibr" rid="ref8">9</xref>
        ] analysed 154,000 tweets that contained the
hashtag ‘#breakingnews”. They determine popularity of messages by counting
retweets and detecting popular terms such as nouns and verbs. This work is
taken further with a simple tf-idf scheme that is used to identify similarity [
        <xref ref-type="bibr" rid="ref9">10</xref>
        ];
named entities are then identified using the Stanford Named Entity Recogniser in
order to identify communities and similar message groups. Sayyadi et al. [
        <xref ref-type="bibr" rid="ref10">11</xref>
        ] also
model the community to discover and detect events on the live Labs SocialStream
platform, extracting keywords, noun phrases and named entities. Ozdikis et al.
[
        <xref ref-type="bibr" rid="ref11">12</xref>
        ] also detect events using hashtags by clustering them and finding semantic
similarities between hashtags, the latter being more of a lexicographic method.
Ratkiewitcz et al. [
        <xref ref-type="bibr" rid="ref12">13</xref>
        ] focus specifically on the detection of a single type of topic,
namely political abuse. Evidence used include the use of hashtags and mentions.
Alvanaki [
        <xref ref-type="bibr" rid="ref13">14</xref>
        ] propose a system based on popular seed tags (tag pairs) which
are then tracked, with any shifts detected and monitored. These articles do use
natural language processing methods and most consider temporal factors, but
do not use n-grams.
      </p>
      <p>
        Becker et al. [
        <xref ref-type="bibr" rid="ref14">15</xref>
        ] also consider temporal issues by focusing on the online
detection of real world events, distinguishing them from non-events (e.g.
conversations between posters). Clustering and classification algorithms are used to
achieve this. Methods such as n-grams and NLP are not considered.
3
3.1
      </p>
      <p>Methods</p>
      <p>
        BNgrams
Term frequency-inverse document frequency, or tf-idf, has been used for indexing
documents since it was first introduced [
        <xref ref-type="bibr" rid="ref15">16</xref>
        ]. We are not interested in indexing
documents however, but in finding novel trends, so we want to find terms that
appear in one time period more than others. We treat temporal windows as
documents and use them to detect words and phrases that are both new and
significant. We therefore define newsworthiness as the combination of novelty and
significance. We can maximise significance by filtering tweets either by keywords
(as in this work) or by following a carefully chosen list of users, and maximise
novelty by finding bursts of suddenly high-frequency words and phrases.
      </p>
      <p>We select terms with a high “temporal document frequency-inverse document
frequency”, or df-idft, by comparing the most recent x messages with the previous
x messages and count how many contain the term. We regard the most recent
x messages as one “slot”. After standard tokenization and stop-word removal,
we index all the terms from these messages. For each term, we calculate the
document frequency for a set of messages using dfti, defined as the number of
messages in a set i that contain the term t.</p>
      <p>df −idfti = (dfti + 1) ·</p>
      <p>1
log dft(i−1) + 1 + 1
.</p>
      <p>(1)</p>
      <p>
        This produces a list of terms which can be ranked by their df-idft scores.
Note that we add one to term counts to avoid problems with dividing by zero or
taking the log of zero. To maintain some word order information, we define terms
as n-grams, i.e. sequences of n words. Based on experiments reported elsewhere
[
        <xref ref-type="bibr" rid="ref2">3</xref>
        ], we use 1-, 2- and 3-grams in this work. High frequency n-grams are likely
to represent semantically coherent phrases. Having found bursts of potentially
newsworthy n-grams, we then group together n-grams that tend to appear in the
same tweets. Each of these clusters defines a topic as a list of n-grams, which we
also illustrate with a representative tweet. We call this process of finding bursty
n-grams“BNgrams.”
3.2
      </p>
      <p>Topic Clustering
An isolated word or phrase is often not very informative, but a group of them
can define the essence of a story. Therefore, we group the most representative
phrases into clusters, each representing a single topic. A group of messages that
discuss the same topic will tend to contain at least some of the same phrases.
We can then find the message that contains the most phrases that define a topic,
and use that message as a human-readable label for the topic. We now discuss
three clustering algorithms that we compare here.</p>
      <p>
        Hierarchical clustering. Here, we initially assign every n-gram to its own
singleton cluster, then follow a standard “group average” hierarchical clustering
algorithm [
        <xref ref-type="bibr" rid="ref16">17</xref>
        ] to iteratively find and merge the closest pair of clusters. We repeat
this until no two clusters share more than half their terms, at which point we
assume that each cluster represents a distinct topic. We define the similarity
between two terms as the fraction of messages in the same time slot that contain
both of them, so it is highly likely that the term clusters whose similarities
are high represent the same topic. Further details about this algorithm and its
parameters can be found in our previous published work [
        <xref ref-type="bibr" rid="ref1">2</xref>
        ].
      </p>
      <p>
        Apriori algorithm. The Apriori algorithm [
        <xref ref-type="bibr" rid="ref17">18</xref>
        ] finds all the associations
between the most representative n-grams based on the number of tweets in which
they co-occur. Each association is a candidate topic at the end of the process.
One of the advantages of this approach is that one n-gram can belong to
different associations (partial membership), avoiding one problem with hierarchical
clustering. No number of associations has to be specified in advance. We also
obtain maximal associations after clustering to avoid large overlaps in the final
set of topic clusters.
      </p>
      <p>
        Gaussian mixture models. GMMs assign probabilities (or strengths) of
membership of each n-gram to each cluster, allowing partial membership of
multiple clusters. This approach does require the number of clusters to be specified
in advance, although this can be automated (e.g. by using Bayesian information
criteria [
        <xref ref-type="bibr" rid="ref18">19</xref>
        ]). Here, we use the Expectation - Maximisation algorithm to optimise
a Gaussian mixture model [
        <xref ref-type="bibr" rid="ref19">20</xref>
        ]. We fix the number of clusters at 20, although
initial experiments showed that using more or fewer produced very similar
results. Seeking more clusters in the data than there are newsworthy topics means
that some clusters will contain irrelevant tweets and outliers, which can later be
assigned a low rank and effectively ignored, leaving us with a few highly-ranked
clusters that are typically newsworthy.
      </p>
    </sec>
    <sec id="sec-2">
      <title>Topic Ranking</title>
      <p>To maximise usability we need to avoid overwhelming the user with a very large
number of topics. We therefore want to rank the results by relevance. Here, we
compare two topic ranking techniques.</p>
      <p>Maximum n -gram df − idft. One method is to rank topics according to the
maximum df − idft value of their constituent n-grams. The motivation of this
approach is assume that the most popular n-gram from each topic represents
the core of the topic.</p>
      <p>Weighted topic-length. As an alternative we propose weighting the
topiclength (i.e. the number of terms found in the topic) by the number of tweets in
the topic to produce a score for each topic. Thus the most detailed and popular
topics are assigned higher rankings. We define this score thus:
st = α · Lt + (1 − α) · Nt (2)</p>
      <p>Lmax Ns
where st is the score of topic t, Lt is the length of the topic, Lmax is the maximum
number of terms in any current topic, Nt is the number of tweets in topic t and
Ns is the number of tweets in the slot. Finally, α is a weighting term. Setting
α to 1 rewards topics with more terms; setting α to 0 rewards topics with more
tweets. We used α = 0.7 in our experiments, giving slightly more weight to those
stories containing more details, although the exact value is not critical.
4</p>
      <p>Experiments
Here, we show the results of our experiments with several variations of the
BNgram approach. We focus on two questions. First, what is best slot size to
balance topic recall and refresh rate? A very small slot size might lead to missed
stories as too few tweets would be analysed; conversely, a very large slot size
means that topics would only be discovered some time after they have
happened. This low ‘refresh rate’ would reduce the timeliness of the results. Second,
what the best combination of clustering and topic ranking techniques? Earlier,
we introduced three clustering methods and two topic ranking methods; we need
to determine which methods are most useful.</p>
      <p>
        We have previously shown that our methods perform well [
        <xref ref-type="bibr" rid="ref1">2</xref>
        ]. The BNgram
approach was compared to a popular baseline system in topic detection and
tracking – Latent Dirichlet Allocation (LDA) [
        <xref ref-type="bibr" rid="ref20">21</xref>
        ] – and to several other
competitive topic detection techniques, getting the best overall topic recall. In addition,
we have shown the benefits of using n-grams compared with single words for this
sort of analysis [
        <xref ref-type="bibr" rid="ref2">3</xref>
        ]. Below, we present and discuss the results from our current
experiments, starting with our approach to evaluation.
4.1
      </p>
    </sec>
    <sec id="sec-3">
      <title>Evaluation Methods</title>
      <p>When evaluating any IR system, it is crucial to define a realistic test problem.
We used three Twitter data sets focused on popular real-world events and
compare the topics that our algorithm finds with an externally-defined ground truth.
To establish this ground truth, we relied on mainstream media (MSM) reports of
the three events. This use of MSM sources helps to ensure that our ground truth
topics are newsworthy (by definition) and that the evaluation is goal-focussed
(i.e. will help journalists write such stories). We filtered Twitter using relevant
keywords and hashtags to collect tweets around three events : the “Super
Tuesday” primaries, part of the presidential nomination race of the US Republican
Party; the 2012 FA Cup final, the climax to the English football season; and the
2012 US presidential election, an event of global significance. In each case, we
reviewed the published MSM accounts of the events and chose a set of stories
that were significant, time-specific, and represented on Twitter. For example, we
ignored general reviews of the state of US politics (not time-specific), and quotes
from members of the public (not significant events).</p>
      <p>For each target topic, we identified around 5-7 keywords that defined the
story to measure recall and precision, as discussed below. Some examples are
shown in the first two columns of Table 4. We also defined several
“forbidden” keywords. A topic was only considered as successfully recalled if all of the
“mandatory” terms were retrieved and none of the “forbidden” terms. The aim
was to avoid producing topics such as “victory Romney Paul Santorum Gingrich
Alaska Georgia” that convey no information about who won or where; or
“Gingrich wins”, which is too limited to define the story because it doesn’t name the
state where the victory occurred.</p>
      <p>
        Figure 1 shows the frequency of tweets collected over time, with further
details in ref. [
        <xref ref-type="bibr" rid="ref1">2</xref>
        ]. We have made all the data freely available3. The three data
sets differ in the rates of tweets, determined by the popularity of the topic and
the choice of filter keywords. The mean tweets per minute (tpm) were: Super
Tuesday, 832 tpm; FA Cup, 1293 tpm; and US elections, 2209 tpm. For a slot
size of 1500 tweets these correspond to a “topic refresh rate” of 108s, 70s and
3 http://www.socialsensor.eu/results/datasets/72-twitter-tdt-dataset
41s respectively. This means that a user interface displaying these topics could
be updated every 1–2 minutes to show the current top-10 (or top-m) stories.
      </p>
      <p>We ran the topic detection algorithm on each data set. This produced a
ranked list of topics, each defined by a set of terms (i.e. n-grams). For our
evaluation, we focus on the recall of the top m topics (1 ≤ m ≤ 10) at the time
each ground-truth story emerges. For example, if a particular story was being
discussed in the mainstream media from 10:00-10:15, then we consider the topic
to be recalled if the system ranked it in the top m at any time during that
period.</p>
      <p>The automatically detected topics were compared to the ground truth
(comprising 22 topics for Super Tuesday; 13 topics for FA Cup final; and 64 topics
for US elections) using three metrics: Topic recall: Percentage of ground truth
topics that were successfully detected by a method. A topic was considered
successfully detected if the automatically produced set of words contained all
mandatory keywords for it (and none of the forbidden terms, if defined).
Keyword precision: Percentage of correctly detected keywords out of the total
number of keywords for all topics detected by the algorithm in the slot.
Keyword recall: Percentage of correctly detected keywords over the total number
of ground truth keywords (excluding forbidden keywords) in the slot. One key
difference between “topic recall” and “keyword recall” is that the former is a
user-centred evaluation metric, as it considers the power of the system at
retrieving and displaying to the user stories that are meaningful and coherent, as
opposed to retrieving only some keywords that are potentially meaningless in
isolation.</p>
      <p>
        Note that we do not attempt to measure topic precision as this would need
an estimate of the total number of newsworthy topics at any given time, in
order to verify which (and how many) of the topics returned by our system
were in fact newsworthy. This would require an exhaustive manual analysis of
MSM sources to identify every possible topic (or some arbitrary subset), which
is infeasible. One option is to compare detected events to some other source,
such as Wikipedia, to verify the significance of the event [
        <xref ref-type="bibr" rid="ref21">22</xref>
        ], but Wikipedia
does not necessarily correspond to particular journalists’ requirements regarding
newsworthiness and does not claim to be complete.
4.2
      </p>
      <p>
        Results
Table 1 shows the effect on topic recall of varying the slot size, with the same
total number of topics in the evaluation for each slot size. The mean is weighted by
the number of topics in the ground truth for each set, giving greater importance
to larger test sets. Overall, using very few tweets produces slightly worse results
than with larger slot sizes (e.g. 1500 tweets), presumably as there is too little
information in such a small collection. Slightly better results for the Super
Tuesday set occur with fewer tweets; this could be due to the slower tweet rate in this
set. Note that previous experiments [
        <xref ref-type="bibr" rid="ref2">3</xref>
        ] showed that including 3-grams improves
recall compared to just using 1- and 2-grams, but adding 4-grams provides no
extra benefit, so here we use 1-, 2- and 3-gram phrases throughout.
      </p>
      <p>Lastly, we compared the results of combining different clustering techniques
with different topic ranking techniques (see Fig. 2). We conclude that the
hierarchical clustering performs well despite the weakness discussed above (i.e.
each n-gram is assigned to only one cluster), especially in FA Cup dataset. Also,
the use of weighted topic-length ranking technique improves topic recall with
hierarchical clustering in the political data sets.</p>
      <p>The Apriori algorithm performs quite well in combination with the weighted
topic length ranking technique (note that this ranking technique was specially
created for the “partial” membership clustering techniques). We see that the
Apriori algorithm in combination with the maximum n-gram df − idft ranking
technique produces slightly worse results, as this ranking technique does not
produce diverse topics for the first results (from top 1 to top 10, in our case) as
we mentioned earlier.</p>
      <p>Turning to the EM Gaussian mixture model results, we see that this method
works very well on the FA Cup final and US elections data sets. Despite being
a “partial” membership clustering technique, the use of weighted topic length
ranking technique does not make any representative difference, even its
performance is worse in Super Tuesday dataset. Further work is needed to test this.</p>
      <p>Table 2 summarises the results of the three clustering methods and the two
ranking methods across all three data sets. The weighted-mean scores show that
for the three clustering methods, ranking by the length of the topic is more
effective than ranking by each topic’s highest df − idft score. We can see that
for the FA Cup set, the Hierarchical and GMM clustering methods are the best
ones in combination with the maximum n-gram df − idft ranking technique.
For Super Tuesday and US Elections data sets, “partial” membership clustering
techniques (Apriori and GMM, respectively) perform the best in combination
with weighted topic length ranking technique, as expected.</p>
      <p>Finally, Table 3 shows more detailed results, including keyword precision and
recall, for the best combinations of clustering and topic ranking methods of the
three datasets when the top five results are considered per slot. In addition,
Table 4 shows some examples of ground truth and BNgram detected topics and
tweets within the corresponding detected topics for all datasets.
0.8
0.6
0.4
0.2</p>
      <p>1
0.8
0.6
0.4
0.2</p>
      <sec id="sec-3-1">
        <title>Super Tuesday FA Cup 1</title>
        <p>0.8
0.6
0.4
0.2
5
(b)</p>
      </sec>
      <sec id="sec-3-2">
        <title>Hierarchical+Max</title>
      </sec>
      <sec id="sec-3-3">
        <title>Apriori+Max</title>
        <p>GMM+Max</p>
      </sec>
      <sec id="sec-3-4">
        <title>Hierarchical+Length</title>
      </sec>
      <sec id="sec-3-5">
        <title>Apriori+Length</title>
      </sec>
      <sec id="sec-3-6">
        <title>GMM+Length</title>
        <p>5
(a)</p>
        <p>US elections
1
3
7
10
1
3
7
10</p>
      </sec>
      <sec id="sec-3-7">
        <title>Number of top results</title>
      </sec>
      <sec id="sec-3-8">
        <title>Number of top results</title>
        <p>1
3
5
7
10</p>
      </sec>
      <sec id="sec-3-9">
        <title>Number of top results (c)</title>
        <p>If we compare the results between the three collections, one difference is
particularly striking: the topic recall is far higher for football (over 90%) than for
politics (around 60-80%; Table 2). This is likely to reflect the different nature
of conversations about the events. Topics within a live sports event tend to be
transient: fans care (or at least tweet) little about what happened five minutes
ago; what matters is what is happening “now”. This is especially true during
key events, such as goals. In politics, conversations and comments tend to spread
over hours (or even days) rather than minutes. This means that sports-related
topics tend to occur over a much narrower window, with less overlapping chatter.
In politics, several different topics are likely to be discussed at the same time,
making this type of trend detection much harder. Looking back at the
distribution of the tweets over time (Figure 1), we can see clear spikes in the FA Cup
graph, each corresponding to a major event (kick-off, goals, half-time, full-time
etc.). No such clarity is in the politics graphs, which instead is best viewed as
many overlapping trends.</p>
        <p>This difference is reflected in the way that major news stories often emerge: an
initial single, focussed story emerges but is later replaced with several potentially
overlapping sub-stories covering different aspects of the story. Our results suggest
that a dynamic approach may be required for newsworthy topic detection, finding
an initial clear burst and subsequently seeking more subtle and overlapping
topics.</p>
        <p>
          Recently, Twitter has been actively increasing its ties to television4.
Broadcast television and sporting events share several common features: they occur a
pre-specified times; they attract large audiences; and they are fast-paced. These
features all allow and encourage audience participation in the form of sharing
comments and holding discussions during the events themselves, such that the
4 “Twitter &amp; TV: Use the power of television to grow your impact” https://
business.twitter.com/twitter-tv
Target topic Ground truth Extracted keywords Example tweet
keywords
Newt Gingrich says Newt Gingrich, launch, March, Mo- @Bailey Shel: RT
@newtgin“Thank you Geor- Thank you, mentum, decisively, grich: Thank you Georgia! It is
gia! It is gratify- Georgia, March, thank, Georgia, gratify- gratifying to win my home state
ing to win my home Momentum, ing, win, home, state, so decisively to launch our March
state so decisively gratifying #MarchMo, #250gas, Momentum. #MarchMo #250gas
to launch our March @newtgingrich
Momentum”
Salomon Kalou has Salomon Kalou, Liverpool, defence, be- @SharkbaitHooHa : RT
an effort at goal run, box, mazy fore, gets, ambushed, @chelseafc: Great mazy run
from outside the Kalou, box, mazy, by Kalou into the box but he
area which goes run, @chelseafc, great, gets ambushed by the Liverpool
wide right of the #cfcwembley, #facup, defence before he can shoot
goal shoot #CFCWembley #FACup
US President Barack Obama, best, America, best, come, @northoaklandnow: “We know
Obama has pledged come United, States, hearts, in our hearts that for the United
”the best is yet to #Obama, speech, know, States of America, the best is yet
come”, following a victory to come,” says #Obama in
vicdecisive re-election tory speech.
victory over
Republican challenger Mitt
Romney
topics extracted by the df-idft algorithm, and example tweets selected by our
system from the collections.
focus of the discussion is constantly moving with the event itself. Potentially,
this can allow targeted time-sensitive promotions and advertising based on
topics currently receiving the most attention. Facebook and other social media are
also competing for access to this potentially valuable “second screen” [
          <xref ref-type="bibr" rid="ref22">23</xref>
          ].
Television shows are increasingly promoting hashtags in advance, which may make
collecting relevant tweets more straightforward. Even if topic detection with
news requires slightly different methods compared to sport and television, both
have substantial and growing demand.
        </p>
        <p>Acknowledgments This work is supported by the SocialSensor FP7 project,
partially funded by the EC under contract number 287975. We wish to thank Nic
Newman and Steve Schifferes of City University London for invaluable advice.</p>
        <p>2011)
1. Newman, N.: Mainstream media and the distribution of news in the age of social
discovery. Reuters Institute for the Study of Journalism working paper (September
1(3) (September 2013)
IEEE Transactions on 15(6) (2013) 1268–1282</p>
        <p>A., Kompatsiaris, I., Jaimes, A.: Sensing trending topics in twitter. Multimedia,
Computer Society Special Technical Community on Social Networking E-Letter</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          2.
          <string-name>
            <surname>Aiello</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Petkos</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Martin</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Corney</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Papadopoulos</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Skraba</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          , Goker,
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          3.
          <string-name>
            <surname>Martin</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Corney</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goker</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Finding newsworthy topics on Twitter</article-title>
          . IEEE
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          4.
          <string-name>
            <surname>Newman</surname>
          </string-name>
          , N.:
          <article-title>#ukelection2010, mainstream media and the role of the internet</article-title>
          .
          <source>Reuters Institute for the Study of Journalism working paper (July</source>
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          5.
          <string-name>
            <surname>Petrovic</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Osborne</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lavrenko</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          :
          <article-title>Streaming first story detection with application to Twitter</article-title>
          .
          <source>In: Proceedings of NAACL</source>
          . Volume
          <volume>10</volume>
          . (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          6.
          <string-name>
            <surname>Petrovic</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Osborne</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lavrenko</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          :
          <article-title>Using paraphrases for improving first story detection in news and Twitter</article-title>
          .
          <source>In: Proceedings of HTL12 Human Language Technologies</source>
          . (
          <year>2012</year>
          )
          <fpage>338</fpage>
          -
          <lpage>346</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          7.
          <string-name>
            <surname>Benhardus</surname>
          </string-name>
          , J.:
          <source>Streaming trend detection in Twitter. National Science Foundation REU for Artificial Intelligence, Natural Language Processing and Information Retrieval</source>
          , University of Colarado (
          <year>2010</year>
          )
          <fpage>1</fpage>
          -
          <lpage>7</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          8.
          <string-name>
            <surname>Shamma</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kennedy</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Churchill</surname>
          </string-name>
          , E.:
          <article-title>Peaks and persistence: modeling the shape of microblog conversations</article-title>
          .
          <source>In: Proceedings of the ACM 2011 conference on Computer supported cooperative work</source>
          ,
          <source>ACM</source>
          (
          <year>2011</year>
          )
          <fpage>355</fpage>
          -
          <lpage>358</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          9.
          <string-name>
            <surname>Phuvipadawat</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Murata</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Breaking news detection and tracking in Twitter</article-title>
          .
          <source>In: Proceedings of the 2010 IEEE/WIC/ACM International Conference on Web Intelligence and Intelligent Agent Technology. Volume</source>
          <volume>3</volume>
          . (
          <year>2010</year>
          )
          <fpage>120</fpage>
          -
          <lpage>123</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          10.
          <string-name>
            <surname>Phuvipadawat</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Murata</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Detecting a multi-level content similarity from microblogs based on community structures and named entities</article-title>
          .
          <source>Journal of Emerging Technologies in Web Intelligence</source>
          <volume>3</volume>
          (
          <issue>1</issue>
          ) (
          <year>2011</year>
          )
          <fpage>11</fpage>
          -
          <lpage>19</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          11.
          <string-name>
            <surname>Sayyadi</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hurst</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Maykov</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Event detection and tracking in social streams</article-title>
          .
          <source>In: Proceedings of International Conference on Weblogs and Social Media (ICWSM)</source>
          .
          <article-title>(</article-title>
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          12.
          <string-name>
            <surname>Ozdikis</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Senkul</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Oguztuzun</surname>
          </string-name>
          , H.:
          <article-title>Semantic expansion of hashtags for enhanced event detection in Twitter</article-title>
          .
          <source>In: Proceedings of VLDB 2012 Workshop on Online Social Systems</source>
          . (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          13.
          <string-name>
            <surname>Ratkiewicz</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Conover</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Meiss</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , Gonc¸alves,
          <string-name>
            <given-names>B.</given-names>
            ,
            <surname>Flammini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Menczer</surname>
          </string-name>
          ,
          <string-name>
            <surname>F.</surname>
          </string-name>
          :
          <article-title>Detecting and tracking political abuse in social media</article-title>
          .
          <source>Proc. of ICWSM</source>
          (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          14.
          <string-name>
            <surname>Alvanaki</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sebastian</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ramamritham</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Weikum</surname>
          </string-name>
          , G.:
          <article-title>Enblogue: emergent topic detection in Web 2.0 streams</article-title>
          .
          <source>In: Proceedings of the 2011 international conference on Management of data, ACM</source>
          (
          <year>2011</year>
          )
          <fpage>1271</fpage>
          -
          <lpage>1274</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          15.
          <string-name>
            <surname>Becker</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Naaman</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gravano</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Beyond trending topics: Real-world event identification on Twitter</article-title>
          .
          <source>In: Proceedings of the Fifth International AAAI Conference on Weblogs and Social Media (ICWSM11)</source>
          . (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          16. Sp¨arck Jones,
          <string-name>
            <surname>K.</surname>
          </string-name>
          :
          <article-title>A statistical interpretation of term specificity and its application in retrieval</article-title>
          .
          <source>Journal of Documentation</source>
          <volume>28</volume>
          (
          <issue>1</issue>
          ) (
          <year>1972</year>
          )
          <fpage>11</fpage>
          -
          <lpage>21</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          17.
          <string-name>
            <surname>Murtagh</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>A survey of recent advances in hierarchical clustering algorithms</article-title>
          .
          <source>The Computer Journal</source>
          <volume>26</volume>
          (
          <issue>4</issue>
          ) (
          <year>1983</year>
          )
          <fpage>354</fpage>
          -
          <lpage>359</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          18.
          <string-name>
            <surname>Agrawal</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Srikant</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          , et al.:
          <article-title>Fast algorithms for mining association rules</article-title>
          .
          <source>In: Proc. 20th Int. Conf. Very Large Data Bases, VLDB</source>
          . Volume
          <volume>1215</volume>
          . (
          <year>1994</year>
          )
          <fpage>487</fpage>
          -
          <lpage>499</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          19.
          <string-name>
            <surname>Fraley</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Raftery</surname>
            ,
            <given-names>A.E.</given-names>
          </string-name>
          :
          <article-title>How many clusters? Which clustering method? Answers via model-based cluster analysis</article-title>
          .
          <source>The Computer Journal</source>
          <volume>41</volume>
          (
          <issue>8</issue>
          ) (
          <year>1998</year>
          )
          <fpage>578</fpage>
          -
          <lpage>588</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          20.
          <string-name>
            <surname>Dempster</surname>
            ,
            <given-names>A.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Laird</surname>
            ,
            <given-names>N.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rubin</surname>
            ,
            <given-names>D.B.</given-names>
          </string-name>
          :
          <article-title>Maximum likelihood from incomplete data via the EM algorithm</article-title>
          .
          <source>Journal of the Royal Statistical Society. Series B (Methodological)</source>
          (
          <year>1977</year>
          )
          <fpage>1</fpage>
          -
          <lpage>38</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          21.
          <string-name>
            <surname>Blei</surname>
            ,
            <given-names>D.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ng</surname>
            ,
            <given-names>A.Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jordan</surname>
            ,
            <given-names>M.I.</given-names>
          </string-name>
          :
          <article-title>Latent Dirichlet Allocation</article-title>
          .
          <source>Journal of Machine Learning Research 3 (Mar</source>
          <year>2003</year>
          )
          <fpage>993</fpage>
          -
          <lpage>1022</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          22.
          <string-name>
            <surname>Osborne</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Petrovic</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>McCreadie</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Macdonald</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ounis</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Bieber no more: First story detection using Twitter and Wikipedia</article-title>
          .
          <source>In: SIGIR 2012 Workshop on Time-aware Information Access</source>
          . (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          23.
          <string-name>
            <surname>Goel</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stelter</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Social networks in a battle for the second screen</article-title>
          .
          <source>The New York Times (October 2</source>
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>