<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Artificial Prediction Markets as a tool for Syndromic Surveillance</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Fatemeh Jahedpari</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Julian Padget</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Marina De Vos</string-name>
          <email>m.d.vosg@bath.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Benjamin Hirsch</string-name>
          <email>Benjamin.Hirsch@kustar.ac.ae</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science, University of Bath</institution>
          ,
          <country country="UK">UK</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>EBTIC, Khalifa University</institution>
          ,
          <country country="AE">United Arab Emirates</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>A range of data sources across the internet, such as google search terms, twitter topics and Facebook messages, amongst others, can be viewed as kinds of sensors from which information might be extractable about trends in the expression of matters of concern to people. We focus on the problem of how to identify emerging trends after the original textual data has been processed into a quantitative form suitable for the application of machine learning techniques. We present some preliminary ideas, including an agent-based implementation and some early results, about the application of artificial prediction markets to such data, taking the specific domain of syndromic surveillance (early stage recognition of epidemics) as an example, using publicly available data sets.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>This paper outlines some early stage research into the application of prediction markets
to syndromic surveillance. Prediction markets are seen as a mechanism to forecast the
outcome of future events by aggregating public opinion, in which market participants
trade so-called securities that represent different probabilities about the (expected)
outcome of a scenario. We describe prediction markets in more detail in section 2 and
compare them with alternative approaches in section 5.</p>
      <p>Syndromic surveillance monitors population health indicators which are apparent
before confirmatory diagnostic tests become available, in order to predict a disease
outbreak within a society at the earliest possible moment, with the aim of protecting
community health. Clearly, the earlier a health threat within a population is detected, the
lower the morbidity and the higher the number of lives that may be saved. Syndromic
surveillance data sources include, but are not limited to, coding of diagnoses at
admission or discharge emergency department, chief complaints, medical encounter
prediagnostic data, absentee rates at schools and workplaces, over-the-counter pharmacy
sales, Internet and open source information such as people post in social media. Each of
these types of data can generate a signal during a disease development. Therefore, given
the vast amount of these data sources, a proper mechanism is necessitated to integrate
them as soon as they become available.</p>
      <p>In this research, we focus on developing a novel syndromic surveillance technique by
integrating different data sources inspired by the crowd-sourcing behaviour of
prediction markets. To achieve our goal, we train a multiagent system in an artificial prediction
market in a semi-supervised manner.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Prediction Markets</title>
      <p>Prediction markets have been used to forecast accurately the outcome of political
contests, sporting events, and economic outcomes [19]. In this research, we use an
artificial prediction market as a mechanism to integrate several syndromic surveillance data
sources to predict a level of disease activity within a population on a specific date. This
section briefly explains the preliminaries of prediction markets.</p>
      <p>
        The prediction market, also known as an information market, originated at the Iowa
Electronic Marketplace (IEM) in 1988 as a means to bet on presidential elections. A
prediction market aims to utilise the aggregated wisdom of the crowd in order to predict
the outcome of a future event [16]. In these markets, traders’ behaviour has the effect of
externalising their private information and beliefs about the possible outcomes, and can
hence be used to forecast an event accurately [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. Prediction markets are increasingly
being considered as approaches for collecting, summarising and aggregating dispersed
information by governments and corporations [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ].
      </p>
      <p>In prediction markets, traders bet on the outcome of future events by trading securities.
A security is a financial instrument, like a financial stock, that pays a profit (or makes
a loss) based on the outcome of the event. Each outcome of an event has a security
associated with it. Traders can buy or sell any number of securities before the expiry
time of the security. A security expires when the outcome of the event is realised. To
illustrate a simple case, a prediction market can be used to predict “if candidate ‘X’ will
win the election” by offering two securities of ‘Yes’ and ‘No’. Assuming the market
finally ends with candidate ‘X’ winning the election, all traders will receive $1 payoff
for each ‘Yes’ security they own and $0 for their ‘No’ securities, losing the money they
spent on buying them.</p>
      <p>The aggregated monetary bets made by market traders dynamically determine the price
of each security before the market ends. The market price of a security represents the
price at which the security can be bought or sold. Also, it can be interpreted as
representing the probability of that outcome occurring by fusing the beliefs of all the market
participants. Arguably, the price that an agent would pay to buy a security indicates
how confident s/he is in the outcome of the event. For example, if a trader believes that
the chance of candidate ‘X’ winning is 80%, s/he then would be willing to buy a ‘Yes’
security at any price up to $0.80.</p>
      <p>
        A prediction market is run by a market-maker who is the company or individual that
interacts with traders to buy and sell securities. The market-maker determines the
market price using a market trading protocol. The logarithmic market scoring rule (LMSR)
designed by Hanson [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] is an automated market maker. Using LMSR, the price and cost
of a security is calculated as follows:
and
m
C(qi) = b log(X eqi=b)
P (qi) =
      </p>
      <p>i=1
exp(qi=b)
Pm</p>
      <p>j=1 exp(qj =b)
respectively, where m is number of securities that market offers, each for one possible
outcome and qi 2 (q1; q2; : : : ; qm) represents the number of units of security i held
by market traders The larger the value of b, the more money the market maker can
lose. It also means that traders can purchase additional quantities of a security without
causing significant price swings. Note that the price of a security only applies for buying
a infinitesimal number of shares and the price of the security immediately changes as
soon as a traders start trading. In order to calculate the cost of a trading X securities,
the market makers must calculate C(q + X) C(q).
3</p>
    </sec>
    <sec id="sec-3">
      <title>Agent-Based Simulation Architecture</title>
      <p>In order to explore empirically the application of artificial prediction markets to
syndromic surveillance, we have developed an agent-based simulation, which we now
describe, followed by some preliminary results in section 4.
3.1</p>
      <sec id="sec-3-1">
        <title>Agents and Strategies</title>
        <p>Our model integrates data and beliefs of different data streams by simulating an artificial
prediction market to predict the outcome of an event, which in this case is the disease
activity level on a specific date. Each data stream includes the quantitative value of a
particular disease activity level for a specific place for different period of time. Each
agent is responsible for one data stream and trades securities in various prediction
markets based on its capital and belief about the disease activity level of the market event
date. Trading agents will in due course (see below) learn from each market based on the
revenue they receive and the losses they make when the market closes. Consequently,
they can update their strategy, beliefs and confidence for the future markets.
The system has a market-maker that uses a scoring rule to calculate the market price
for each security in the market, and a data distributor that provide agents with the data
stream for which they are responsible, and trading agents. The simulation mechanism,
specified in Algorithm 1, is as follows. At the beginning of the simulation, all the trading
agents are awarded an equal amount of initial capital. For each training example, a
prediction market (let us say prediction market for week T ) is established. At this time,
the Data Distributor Agent provides available data to the trading agents, according to
their role. Then, the trading agents participate in the market according to their available
capital, beliefs, and trading strategies. Agents can trade any number of securities before
the deadline for the closing of the market. Once the market deadline is reached, the
market-maker reveals the winning security and rewards the winning security holders
with $1 for each winning security they own. These revenues are added to their capital.
However, the agents who own losing securities, lose capital equal to the amount spent
on purchasing them.</p>
        <p>During the simulation, agents with superior data, strategy and analysis algorithms are
likely to accumulate greater capital and hence affect market prices and eventually the
outcome. In other words, important – by these metrics – agents are identifiable and have
greater influence in predicting the outcome of the event. This increased influence of the
more successful agents should increase the accuracy of the system overall: the agents
are not in competition per se, so we do not care which agents are better, but we do want
the better ones to have more effect on the prediction mechanism.</p>
        <p>
          The first agent strategy is based on zero intelligence [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ], and so has no scope for
learning: buy-sell and security choices are random subject to the constraint of not trading at a
loss. For the second strategy, we add a basic learning mechanism, following the design
of zero intelligence plus [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ], in which agents update their trading strategy and beliefs
based on the reward they received from the market in order to improve their reward in
future markets. This is achieved by incorporates a simple machine learning mechanism
(Widrow-Hoff) to adapt their individual behaviour to the market trend.
3.2
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>Market Instantiation</title>
        <p>Different data sources have different timeliness in detecting a disease outbreak. For
example, some data sources such as social media data can signal a disease activity level
perhaps two weeks earlier than physician data. Therefore, for the system to be capable
of forecasting the outcome at the earliest possible moment and not wait for all the
agents’ data to arrive to start prediction, we will run multiple concurrent markets for
consecutive prediction weeks. For example, if the simulation week number is 1, then
4 further markets for weeks 2 to 5 will also be open. Once the deadline of the first
market (week 2) is reached, then that market (week 2) closes and another market after
the last market is opened (week 6). Consequently, the agents who have data for those
further markets can start trading earlier in those markets and take advantage of cheaper
prices, which will lead to the updating of market prices as early as possible. In addition,
all data with different timeliness ranging from 4 weeks to one day before the event date
will be incorporated in each market and at the same time agents can use their knowledge
achieved from the previous market when predicting the outcome of a given market.</p>
        <p>Algorithm 1: Agent-Based Simulation Architecture Algorithm
1 Give start up capital to each agent
For the sake of simplicity, Algorithm 1 considers just the one market, as does our current
implementation.
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Preliminary Results</title>
      <p>The configuration of the controlling parameters of our system need thorough
investigation through running a large number of simulation experiments. These settings include
but are not limited to constraints on number of market participants, required time for
each market, initial capital for each agent, type of monitored diseases, constraints and
requirements for agents to trade, and the minimum required number of training
examples. We have only just begun to explore this parameter space.</p>
      <p>
        As discussed in Section 1, there is a vast number of syndromic surveillance data sources.
Much research has been done to compare these data sources with the actual value of
disease activity level for a specific disease in a particular place. For example, Culotta [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]
stated that he could track influenza rates in the United States using Twitter messages
with 95% correlation. Corley [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] could track flu rates in the United States with a
correlation of 76% by examining the proportion of blogs containing the two keywords of
“influenza” and “flu”. Google Flu Trend [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] can predict ’flu activity level with a 97%
correlation by analysing queries sent to the Google search engine.
      </p>
      <p>
        In the first two batches of experiments, we have tried two well-known trading strategies
for trading agents: Zero Intelligence (ZI) [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] and Zero Intelligence Plus (ZIP) [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], the
former to provide a baseline behaviour and the latter to investigate the effect of a simple
learning mechanism on the trading decision. We now discuss each of these in more
detail.
4.1
      </p>
      <sec id="sec-4-1">
        <title>Preliminary Results with ZI</title>
        <p>
          In the first set of experiments, agents use a Zero Intelligence strategy (adapted from
[
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]) when trading securities in the market. In this model, agents consider a limit price,
according to their data, for each security of the market. In each day of the market, they
choose one security randomly and purchase a random quantity of that security, if its
limit price is higher than its market price or they sell a random quantity of that security
if its limit price is less than its market price. In both situations, the agents considers the
maximum number of securities that can be traded, based on their available capital and
the securities they own. LMSR, described in Section 2, is used as market scoring rule,
as it provides infinite liquidity [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] and does not suffer in thin markets where the number
of traders are small.
        </p>
        <p>Each agent in the experiment is awarded $10 at start up and one market is established
for each training example. The winning security is chosen based on the United State
influenza-like illnesses rate from 30 September 2002 to 01 September 2003 3. Each
market offers eight securities, corresponding to one security for each standard deviation
from the mean, covering from 4 to +4 standard deviations.</p>
        <p>Figure 1 shows how extending the period of a prediction market can help agents to
predict the outcome of the event better. It demonstrates that accuracy goes up and the
mis-classification rate falls as the duration of the market increases. As can be seen from
the figure, accuracy increased from 82% for 10-day-long markets to 93% for
90-daylong markets, after which it is almost flat. From this, we conclude that 110 days seems a
sufficient period for each market and hence our subsequent experiments at this stage use
this market length. Longer duration markets provide agents with sufficient time to trade
enough numbers of the desired securities and approach more closely the equilibrium
price.</p>
        <p>The purpose of the experiments reported in this paper is primarily to establish
confidence in the behaviour of the simulation, by providing tailored data feeds with known
3 http://www.cdc.gov/flu/weekly/fluviewinteractive.htm
properties and then observing whether the agents achieve their expected level of
performance, given that data and their (known) strategy. These experiments have a population
of 20 agents, each receiving data from data streams with a specified correlation with
the United States influenza-like illnesses rate. The agent names in the following figures
represents the type of data the agent is receiving. For instance, a95 denotes an agent
that receives data with a 95% correlation with the United States influenza-like illnesses
rate.</p>
        <p>Since all the agents in these experiments are essentially identical in terms of strategy,
the difference in their data sources should lead to them obtaining different amounts
of revenue in each market. Figure 2 shows their revenue when each market ends. As
the Figure shows, a100 agent, which has complete information about all the events of
the experiment, earns high revenue in all markets with the exception of 31/07/2003. In
this case, it was not making the wrong choice – it cannot – but was trying to purchase
a large number of securities and since the price of a security increases as a result of
its purchase, the agent (as explained in Section 2), did not have sufficient capital to
complete the deal. In other words, for agents without perfect information, the enforced
random choice of security to trade (the ZI strategy) means the agent cannot select the
most appropriate one, but rather the one that chance dictates and hence it makes a loss.
Figure 3 shows the capital held by each agent at the end of each market in one
experiment and Figure 4 shows the average capital of agents over 50 runs. The main
observation from these figures is that, as expected, agents with higher quality data are
able to achieve higher levels of revenue. As can be seen from these figures, agent a100
accumulates more capital than other agents even those with high quality data such as
a99. The reason for this is that a100 never makes a mistake while the other agents do
and as soon as one agent predicts an outcome, it dedicates most of its capital to purchase
the corresponding security. Therefore, once an agent predicts a wrong outcome, it loses
all its capital, while agent a100 keeps earning revenue in each market and accumulates
more capital and hence invests more on upcoming markets and earns more revenue
again. Also, agent a100 causes the price of the correct security to increase rapidly as it
purchase a large quantity of it and, therefore, makes it difficult for other agents to buy
significant quantities of that security due its high price.</p>
        <p>
          Clearly, more comprehensive experimentation is necessary, backed up with appropriate
statistical confidence tests. In this section, we have only used the most basic of strategies
and one that has known flaws [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ]. However, it provides both a useful baseline
performance, as well as a setting in which initial hypotheses about the effectiveness of the
prediction market model can be validated (such as the agent with 100% correlated data
dominating the market and all others losing all their investments).
4.2
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>Preliminary Results with ZIP</title>
        <p>In the second set of experiments, we have changed the agent strategy from ZI to ZIP, by
which the agent uses data about trends in the market in order to adjust their behaviour to
Fig. 1. ZI: Comparing
accuracy (s.d.: 0.017–0.047)
and mis-classification (s.d.:
0.101–0.289) on the y-axis
vs. duration of the prediction
market (x-axis). Each data
point is the average of 50
experiments with same
parameter settings
be less random (as in ZI) and more in line with the market valuation of a given security.
Although ZIP and its variants have been shown to be effective strategies in terms of
profit making, there are two reasons why this approach is likely to be ineffective in the
context of prediction markets:
1. The ZIP strategy depends upon both buyer and seller employing this strategy, but in
the prediction market, the two parties are the buyer and the market-maker, of which
the latter has no interest in profit and which has no strategy as such. Consequently,
only one party in the market is ‘learning’. This still has a positive effect as discussed
below, but starts to underline the difference between trading markets (with bilateral
strategies) and prediction markets (with unilateral strategies).
2. The ZIP strategy aims at trading for profit regardless of the (financial) instrument
being traded, leading to the establishment of an equilibrium price, whereas the point
of a prediction market is to choose the right instrument, rather than the currently
most profitable.</p>
        <p>The experiments are run with the same data as for ZI. Thus Figures 6 and 7 show the
results from a randomly chosen 110 day market, as was done for ZI. It is notable that
ZIP achieved higher accuracy – nearly 97% – with market durations of &gt; 30 days, than
in the ZI experiment. However, as before a100 agent dominates the market.
The story in terms of revenue (Figure 6) is much the same as for ZI, although a99
stops making a profit much sooner with ZIP. As can be seen from the figure, the agent
obtains revenue until 04/11/2002, which is the first time that it makes a mistake and
loses the majority of its capital. Consequently, the agent has little remaining capital,
but continues earning money for the following two markets, but its second mistake (on
25/11/2002) bankrupts it, after which it cannot invest further. This scenario applies to
all other agents and causes a100 to dominate the market, as it never makes any mistake.
5</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Related Work</title>
      <p>
        Many syndromic surveillance systems exist worldwide, each designed for a specific
country, region or state [
        <xref ref-type="bibr" rid="ref11 ref13">11, 20, 15, 13, 17</xref>
        ]. We refer to them as traditional, since they
do not utilise internet based data. While these systems can detect an outbreak with high
accuracy, they suffer from slow response times. For example, the Centers for Disease
Control and Prevention (CDC) publishes USA national and regional data typically with
a 1–2 week reporting lag. It monitors over 3,000 health providers nationwide to report
the proportion of patients seen that exhibit influenza-like illnesses (ILI) 4 [
        <xref ref-type="bibr" rid="ref5 ref6">5, 6</xref>
        ].
On the other hand, modern syndromic surveillance systems appeal to internet based
data such as search engine queries, health news, and peoples’ posts on social networks
4 http://www.cdc.gov/flu/weekly/fluactivity.htm
to predict an outbreak earlier [
        <xref ref-type="bibr" rid="ref2 ref4">18, 2, 4</xref>
        ], albeit with necessarily lower precision. While
some of them claim that they could achieve high accuracy, they are vulnerable to false
alarms [
        <xref ref-type="bibr" rid="ref1 ref6">6, 1</xref>
        ] due to their dependence on a single data stream and disregarding the
benefits from fusing different data sources. Ginsberg et al [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] state, regarding Google Flu
Trends, that “Despite strong historical correlations, our system remains susceptible to
false alerts caused by a sudden increase in ILI-related queries. An unusual event, such
as a drug recall for a popular cold or flu remedy, could cause such a false alert”.
To the best of our knowledge, there is no system that fuses both traditional and internet
based data sources. This could be due to the different timescales that these data sources
have and the consequent issues of appropriate synchronisation.Prediction markets can
overcome this problem as traders can trade securities as soon as they receive new
information and impact the price and consequent probability of an event outcome. It is
interesting to note that Polgreen et al [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] report on the use a prediction market with
human health care expert participants to forecast infectious disease activity 2-4 weeks
in advance.
      </p>
      <p>Moreover, internet based system are only suitable for places where sufficient source
data is available. For example, twitter-based systems cannot have a high accuracy on
places where using twitter is not very common, if even accessible. In addition, even if
sufficient data is available, system accuracy cannot be guaranteed worldwide since
peoples’ behaviour changes from place to place, reflecting differing (digital) cultures. For
example, people in a particular city may seek a physician as soon as they encounter the
symptoms of a disease and do not trust online information, while people in another city
may defer visiting a doctor and seek out online information in order to cure themselves
at the early stages of their sickness. Furthermore, peoples’ behaviour may change over
time. For example, a particular social media may become less popular and cede its role
to newer technology over the time.
6</p>
    </sec>
    <sec id="sec-6">
      <title>Discussion</title>
      <p>Since we are in an early stage of this research, a substantial part of the work is to come.
We have numerous ideas that have yet to be implemented, including: (i) the learning
capability of agents, (ii) consideration of the confidence of agents, (iii) of the different
timeliness of data streams, and (iv) the effect of a heterogeneous population of agents
with different trading strategy and risk prediction model –among other characteristics.
The very preliminary results we have meet our broad expectations for the behaviour of
prediction markets, but it is too early to say whether they can be a general-purpose tool
with useful levels of precision and recall across a range of domains. The ZI strategy,
being essentially random under the constraint of not making a loss, establishes a
useful performance baseline, as well as a framework against which to validate the basic
system hypotheses. The ZIP strategy, while appropriate for bilateral markets seeking to
establish equilibrium prices, is inappropriate – at least, as conventionally formulated –
for prediction markets, although the dampening effect of the learning mechanism does
lead to higher prediction rates and smoother overall behaviour.</p>
      <p>We welcome feedback on the appropriateness of the approach and the above directions
for development as well as alternative mechanisms that might be incorporated in the
prediction market setting.
15. Emergency department syndromic surveillance system: England &amp; Northern Ireland,</p>
      <p>September 2013.
16. Russ Ray. Prediction markets and the financial “wisdom of crowds”. Journal of Behavioral</p>
      <p>Finance, 7(1):2–4, 2006.
17. Communicable and respiratory disease report for england &amp; wales, October 2013.</p>
      <p>Available from: http://www.hpa.org.uk/webc/HPAwebFile/HPAweb_C/
1317139997178, retrieved 20131216.
18. Alessio Signorini, Alberto Maria Segre, and Philip M. Polgreen. The use of twitter to track
levels of disease activity and public concern in the U.S. during the Influenza A H1N1
Pandemic. PLoS ONE, 6:e19467, 05 2011.
19. Erik Snowberg, Justin Wolfers, and Eric Zitzewitz. Prediction markets for economic
forecasting. Working Paper 18222, National Bureau of Economic Research, July 2012. Available
from: http://www.nber.org/papers/w18222.
20. Fu-Chiang Tsui, Jeremy U Espino, Virginia M Dato, Per H Gesteland, Judith Hutman, and
Michael M Wagner. Technical description of rods: a real-time public health surveillance
system. Journal of the American Medical Informatics Association, 10(5):399–408, 2003.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <given-names>Harshavardhan</given-names>
            <surname>Achrekar</surname>
          </string-name>
          , Avinash Gandhe, Ross Lazarus,
          <string-name>
            <surname>Ssu-Hsin Yu</surname>
          </string-name>
          , and Benyuan Liu.
          <article-title>Predicting flu trends using twitter data</article-title>
          .
          <source>In Computer Communications Workshops (INFOCOM WKSHPS)</source>
          ,
          <source>2011 IEEE Conference on</source>
          , pages
          <fpage>702</fpage>
          -
          <lpage>707</lpage>
          . IEEE,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <given-names>Herman</given-names>
            <surname>Anthony</surname>
          </string-name>
          Carneiro and
          <string-name>
            <given-names>Eleftherios</given-names>
            <surname>Mylonakis</surname>
          </string-name>
          .
          <article-title>Google trends: a web-based tool for real-time surveillance of disease outbreaks</article-title>
          .
          <source>Clinical infectious diseases</source>
          ,
          <volume>49</volume>
          (
          <issue>10</issue>
          ):
          <fpage>1557</fpage>
          -
          <lpage>1564</lpage>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <given-names>Dave</given-names>
            <surname>Cliff</surname>
          </string-name>
          and
          <string-name>
            <given-names>Janet</given-names>
            <surname>Bruten</surname>
          </string-name>
          .
          <article-title>Minimal-intelligence agents for bargaining behaviours in market-based environments</article-title>
          .
          <source>Technical Report HPL-97-91</source>
          ,
          <string-name>
            <surname>Hewlett-Packard Laboratories</surname>
          </string-name>
          ,
          <year>1997</year>
          . Available via http://www.hpl.hp.com/techreports/97/HPL-97-91. html, retrieved
          <volume>20071115</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Courtney D Corley</surname>
          </string-name>
          , Diane J Cook, Armin R Mikler, and Karan P Singh.
          <article-title>Text and structural data mining of influenza mentions in web and social media</article-title>
          .
          <source>International journal of environmental research and public health</source>
          ,
          <volume>7</volume>
          (
          <issue>2</issue>
          ):
          <fpage>596</fpage>
          -
          <lpage>615</lpage>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <given-names>Aron</given-names>
            <surname>Culotta</surname>
          </string-name>
          .
          <article-title>Detecting influenza outbreaks by analyzing twitter messages</article-title>
          .
          <source>arXiv preprint arXiv:1007.4748</source>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <given-names>Jeremy</given-names>
            <surname>Ginsberg</surname>
          </string-name>
          , Matthew H Mohebbi,
          <article-title>Rajan S Patel, Lynnette Brammer, Mark S Smolinski,</article-title>
          and
          <string-name>
            <given-names>Larry</given-names>
            <surname>Brilliant</surname>
          </string-name>
          .
          <article-title>Detecting influenza epidemics using search engine query data</article-title>
          .
          <source>Nature</source>
          ,
          <volume>457</volume>
          (
          <issue>7232</issue>
          ):
          <fpage>1012</fpage>
          -
          <lpage>1014</lpage>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Dhananjay</surname>
            <given-names>K</given-names>
          </string-name>
          <string-name>
            <surname>Gode and Shyam Sunder</surname>
          </string-name>
          .
          <article-title>Allocative efficiency of markets with zerointelligence traders: Market as a partial substitute for individual rationality</article-title>
          .
          <source>Journal of political economy</source>
          , pages
          <fpage>119</fpage>
          -
          <lpage>137</lpage>
          ,
          <year>1993</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <given-names>Robin</given-names>
            <surname>Hanson</surname>
          </string-name>
          .
          <article-title>Combinatorial information market design</article-title>
          .
          <source>Information Systems Frontiers</source>
          ,
          <volume>5</volume>
          (
          <issue>1</issue>
          ):
          <fpage>107</fpage>
          -
          <lpage>119</lpage>
          ,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <given-names>Robin</given-names>
            <surname>Hanson</surname>
          </string-name>
          .
          <article-title>Logarithmic market scoring rules for modular combinatorial information aggregation</article-title>
          .
          <source>Journal of Prediction Markets</source>
          ,
          <volume>1</volume>
          :
          <fpage>3</fpage>
          -
          <lpage>15</lpage>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Robin</surname>
            <given-names>Hanson</given-names>
          </string-name>
          , Ryan Oprea, and
          <string-name>
            <given-names>David</given-names>
            <surname>Porter</surname>
          </string-name>
          .
          <article-title>Information aggregation and manipulation in an experimental market</article-title>
          .
          <source>Journal of Economic Behavior &amp; Organization</source>
          ,
          <volume>60</volume>
          (
          <issue>4</issue>
          ):
          <fpage>449</fpage>
          -
          <lpage>459</lpage>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Michael D Lewis</surname>
          </string-name>
          , Julie A Pavlin, Jay L Mansfield,
          <string-name>
            <surname>Sheilah</surname>
            <given-names>OBrien</given-names>
          </string-name>
          , Louis G Boomsma,
          <article-title>Yevgeniy Elbert</article-title>
          , and Patrick W Kelley.
          <article-title>Disease outbreak detection system using syndromic data in the greater Washington DC area</article-title>
          .
          <source>American journal of preventive medicine</source>
          ,
          <volume>23</volume>
          (
          <issue>3</issue>
          ):
          <fpage>180</fpage>
          -
          <lpage>186</lpage>
          ,
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <given-names>Evdokia</given-names>
            <surname>Nikolova</surname>
          </string-name>
          and
          <string-name>
            <given-names>Rahul</given-names>
            <surname>Sami</surname>
          </string-name>
          .
          <article-title>A strategic model for information markets</article-title>
          .
          <source>In Proceedings of the 8th ACM conference on Electronic commerce</source>
          , pages
          <fpage>316</fpage>
          -
          <lpage>325</lpage>
          . ACM,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <given-names>Yasushi</given-names>
            <surname>Ohkusa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M</given-names>
            <surname>Shigematsu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K</given-names>
            <surname>Taniguchi</surname>
          </string-name>
          , and
          <string-name>
            <given-names>N</given-names>
            <surname>Okabe</surname>
          </string-name>
          .
          <article-title>Experimental surveillance using data on sales of over-the-counter medications-japan,</article-title>
          <string-name>
            <surname>November</surname>
          </string-name>
          2003-April
          <year>2004</year>
          .
          <source>MMWR Morb Mortal Wkly Rep</source>
          ,
          <volume>54</volume>
          :
          <fpage>47</fpage>
          -
          <lpage>52</lpage>
          ,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Philip M. Polgreen</surname>
            ,
            <given-names>Forrest D.</given-names>
          </string-name>
          <string-name>
            <surname>Nelson</surname>
          </string-name>
          , and
          <string-name>
            <surname>George</surname>
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Neumann</surname>
          </string-name>
          .
          <article-title>Use of prediction markets to forecast infectious disease activity</article-title>
          .
          <source>Clinical Infectious Diseases</source>
          ,
          <volume>44</volume>
          (
          <issue>2</issue>
          ):pp.
          <fpage>272</fpage>
          -
          <lpage>279</lpage>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>