<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Performance evaluation of large-scale Information Retrieval systems scaling down</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Fidel Cacheda</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Víctor Carneiro</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Diego Fernández</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Vreixo Formoso</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Facultad de Informática Campus de Elviña</institution>
          <addr-line>s/n 15071, A Coruña</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
      </contrib-group>
      <fpage>36</fpage>
      <lpage>39</lpage>
      <abstract>
        <p>The performance evaluation of an IR system is a key point in the development of any search engine, and specially in the Web. In order to get the performance we are used to, Web search engines are based on large-scale distributed systems and to optimise its performance is an important aspect in the literature. The main methods, that can be found in the literature, to analyse the performance of a distributed IR system are: the use of an analytical model, a simulation model and a real search engine. When using an analytical or simulation model some details could be missing and this will produce some differences between the real and estimated performance. When using a real system, the results obtained will be more precise but the resources required to build a large-scale search engine are excessive. In this paper we propose to study the performance by building a scaled-down version of a search engine using virtualization tools to create a realistic distributed system. Scaling-down a distributed IR system will maintain the behaviour of the whole system and, at the same time, the computer requirements will be softened. This allows the use of virtualization tools to build a large-scale distributed system using just a small cluster of computers.</p>
      </abstract>
      <kwd-group>
        <kwd>Distributed Information Retrieval</kwd>
        <kwd>Performance evaluation</kwd>
        <kwd>Scalability</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Categories and Subject Descriptors</title>
      <p>H.3.4 [Information Storage and Retrieval]: Systems
and Software—Distributed systems; H.3.4 [Information
Storage and Retrieval]: Systems and Software—Performance
evaluation (efficiency and effectiveness)
Copyright c 2010 for the individual papers by the papers’ authors.
Copying permitted only for private and academic purposes. This volume is
published and copyrighted by its editors.</p>
      <p>LSDS-IR Workshop, July 2010. Geneva, Switzerland.</p>
    </sec>
    <sec id="sec-2">
      <title>1. INTRODUCTION</title>
      <p>Web search engines have changed our perspective of the
search process because now we consider normal being able to
search through billions of documents in less than a second.
For example, we may not quite understand why we have to
wait so long in our council for a certificate as they just have
to search through a ”few” thousand/million records.</p>
      <p>However, Web search engines have to use a lot of
computational power to get the performance we are used to.
This computational power can only be achieved using
largescale distributed architectures. Therefore, it is extremely
important to determine the distributed architectures and
techniques that allow clear improvements in the system
performance.</p>
      <p>The performance of a Web search engine is determined
basically by two factors:
• Response time: the time it takes to answer the query.</p>
      <p>This time includes the network transfer times that, in
Internet, will take a few hundred milliseconds; and the
processing time in the search engine, that is usually
limited to 100 milliseconds.
• Throughput: the number of queries the search engine
is able to process per second. This measure usually
has to maintain a constant ratio, but also deals with
peak loads.</p>
      <p>From the user’s point of view, only the response time is
visible and it is the main factor, keeping constant the quality
of the results: the faster the search engine is able to answer
the better. From the search engine point of view both
measures are important. Once an upper limit has been set up
for the response time (e.g. a query should be answered in
less than 100 milliseconds), the objective is to maximise the
throughput.</p>
      <p>
        From the search engine point of view another two factors
have to be taken into account:
• Size: the number of documents indexed by the search
engine. Not so long ago, Google published in its main
page the number of Web pages indexed. Nowadays, the
main commercial search engines do not make public
detailed figures, although the estimations are in the
order of 20 billion documents.
• Resources: the number of computers used by the search
engine. This could be considered from the economical
perspective as the cost of the distributed system. In [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]
Baeza-Yates et al. estimate that, a search engine will
need about 30 thousand computers to index 20 billion
documents and obtain a good performance.
      </p>
      <p>If we want to compare different distributed indexing
models or test some new techniques for a distributed search
engine (e.g. a new cache policy), usually we will fix the size
of the collection and the resources and then, measure the
performance in terms of response time and throughput.</p>
      <p>Ideally, we would need a replica of a large-scale IR
system (for example, one petabyte of data and one thousand
computers) to measure performance. However, this would
be extremely expensive and no research group, or even
commercial search engine, can devote such amount of resources
only to evaluation purposes.</p>
      <p>In this article we present a new approach for performance
evaluation of large-scale IR systems based on scaling-down.
We consider that, creating a scaled-down version of an IR
system, will produce valid results for the performance
analysis, using very few resources. This is an important point
for commercial search engines (from the economical point
of view), but it is more important for the research groups
because this could open the experimentation on large-scale
IR to nearly any group.</p>
      <p>The rest of the paper is organised as follows. In Section 2
we present the main approaches for performance evaluation.
Section 3 analyses our proposal and Section 4 concludes this
work and describes some ideas for future works.</p>
    </sec>
    <sec id="sec-3">
      <title>PERFORMANCE EVALUATION</title>
      <p>In the literature there are many articles that evaluate the
performance of a search engine or one of its components. We
do not intend to present an exhaustive list of papers about
performance evaluation but to present the main methods
used by the researchers, specially of a large-scale IR system.</p>
      <p>The main methods to test the performance of a search
engine are the following:
• An analytical model.
• A simulation model.
• A real system or part of a real system.</p>
      <p>
        A clear example of study based on an analytical model
can be found in [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. In this work, Chowdhury et al. use
the queueing network theory to model a distributed search
engine. The authors model that the processing time in a
query server is a function of the number of documents
indexed. They build a framework in order to analyse
distributed architectures for search engines in terms of response
time, throughput and utilization. To show the utility of this
framework, they provide a set of requirements and study
different scalability strategies.
      </p>
      <p>
        There are many works based on simulation that study the
performance of a distributed IR system. [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] is one of the first.
In this work, Burkowski uses a simple simulation model to
estimate the response time of a distributed search engine,
and uses one server to estimate the values for the simulation
model (e.g. the reading time from disk is approximated as
a Normal distribution). Then, the simulation model
represents a clusters of servers and estimates the response times
using local index organisation (named uniform distribution).
However, the network times are not considered in the
simulation model.
      </p>
      <p>
        Tomasic and Garcia-Molina [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] also used a simulation
model to study the performance of several parallel query
processing strategies using various options for the
organization of the inverted index. They use different simulation
models to represent the collection documents, the queries,
the answer set and the inverted lists.
      </p>
      <p>
        Cacheda et al. in [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] include also a network model to
simulate the behaviour of the network in a distributed IR
system. They compare different distribution architectures
(global and local indexing), identify the main problems and
present some specific solutions, such as, the use of partial
result sets or the hierarchical distribution for the brokers.
      </p>
      <p>
        Other authors use a combination of both approaches. For
example, in [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], Ribeiro-Neto and Barbosa use a simple
analytical model to estimate the processing time in a
distributed system. This analytical model calculates the seek
time for a disk, the reading time from disk of an inverted list,
the time to compare and swap two terms and the transfer
time from one computer to another. In their work, they
include a small simulator to represent the interference among
the various queries in a distributed environment. They
compare the performance of a global index and a local index and
study the effect of the network and disk speed.
      </p>
      <p>
        Some examples of works experimenting with a real IR
system could be [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] or [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. In the first work, Badue et al. study
the imbalance of the workload in a distributed search engine.
They use a configuration of 7 index servers and one broker
to index a collection of 10 million Web pages. In their work,
the use of a real system for testing was important to detect
some important factors for imbalance in the index servers.
They state that the correlations between the term frequency
in a query log and the size of its inverted list lead to
imbalances in query execution times, because these correlations
affect the behaviour of the disk caching.
      </p>
      <p>
        Moffat et al. in [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] study a distributed indexing technique
named pipelined distribution. In a system of 8 servers and
one broker, they index the TREC Terabyte collection [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]
to run their experiments. The authors compare three
distributed architectures: local indexing (or document
partitioning), global indexing (or term partitioning) and
pipelining. In their experiments the pipelined distribution
outperforms the term partitioning, but not the document
partitioning due to a poor workload balancing. However, they also
detect some advantages over the document distribution: a
better use of memory and fewer disk seeks and transfers.
      </p>
      <p>The main drawback for an analytical model is that it
cannot represent all the characteristics of a real IR system.
Some features have to be dropped to keep the model simple
and easy to implement.</p>
      <p>Using a simulation model, we can represent more complex
behaviours than an analytical model. For example, instead
of assuming a fixed transfer time for the network, we can
simulate the behaviour of the network (e.g. we could detect
a network saturation). But, again, not all the features of a
real system could be implemented. Otherwise, we will end
up with a real IR system and not a simulation model.</p>
      <p>In both cases, it is important to use a real system to
estimate the initial values of the model (analytical or simulated)
and, in fact, this is a common practise in all the research
works. In a second step, it is also common to compare the
results of the model with the response obtained from a real
system, using a different configuration, in order to validate
the model.</p>
      <p>However, when the models are used to extrapolate the
behaviour of a distributed IR system, for example increasing
the number computers, the results obtained may introduce
a bigger error than expected. For example, a simulation
model of one computer, when compared with a real system,
has an accuracy of 99% (or an error of 1%). But, what is the
expected error when simulating a system with 10 computers:
1% or 10%?</p>
      <p>This problem is solved by using a real system for the
performance evaluation. But, in this case, the experiments will
be limited by the resources available in the research groups.
In fact, many researchers run their experiments using 10-20
computers. Considering the size of data collections and the
size of commercial search engines, this could not be enough
to provide interesting results for the research community. In
this sense, the analytical or simulation models allow us to
go further, at the risk of increasing the error ratio of our
estimations.</p>
    </sec>
    <sec id="sec-4">
      <title>OUR PROPOSAL</title>
      <p>In this article we propose to use a scaled-down version of
a search engine to analyse the performance of a large-scale
search engine.</p>
      <p>
        Scaling down has been successfully applied in many other
disciplines, and it is specially interesting when the
development of a real system is extremely expensive. For example,
in the shipping industry the use of scale models in basins
is an important way to quantify and demonstrate the
behaviour of a ship or structure, before building a real ship
[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>
        The use of a wind tunnel is also quite common in the
aeronautical or car industries. Specially in the former, the
scaled-down models of planes or parts of a plane are
important to analyse the performance of the structure [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. Also
in architecture scaled-down models are used to test and
improve the efficiency of a building [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
      </p>
      <p>In the world of search engines, is it possible to build a
scaled-down version of a search engine?</p>
      <p>Let us say that we want to study the performance of a
large-scale search engine, composed of 1000 computers, with
the following parameters:
• The size of the collection is 1 petabyte.
• Each computer has a memory of 10 gigabytes.
• Each computer has a disk of 1 terabyte.
• The computers are interconnected using a high speed
network (10Gbits/second).</p>
      <p>From our point of view, maintaining the 1000 computers
as the core of the distributed system, if we apply a scale
factor of 1:1000 we will have a scaled-down version of the
search engine with the following parameters:
• The size of the collection is 1 gigabyte.
• Each computer has a memory of 10 megabytes.
• Each computer has a disk of 1 gigabyte.
• The computers are interconnected using a high speed
network (10Mbits/second).</p>
      <p>One important point is that the scale factor does not
apply to the number of computers. The computers constitute
the core of the distributed system and therefore cannot be
diminished, instead they are scaled-down. This is equivalent
to build a scaled-down version of a building: the beams are
scaled-down but not diminished.</p>
      <p>In this way, we expect to obtain a smaller version of the
large-scale search engine, but with the same drawbacks and
benefits.</p>
      <p>The next step is how to build this scaled-down version of
a search engine.</p>
      <p>The first and trivial solution is to use 1000 computers with
the requirements stated previously. These would be very
basic computers nowadays but, anyway, it could be quite
complicated to obtain 1000 computers for a typical research
group. It could be a little bit easier for a commercial search
engine if they could have access to obsolete computers from
previous distributed systems, but it is not straightforward.</p>
      <p>
        A more interesting solution would be to use virtualization
tools to create the cluster of computers. Virtualization is a
technology that uses computing resources to present one or
many operating systems to user. This technology is based on
methodologies like hardware and software partitioning,
partial or complete machine simulation and others [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. In this
work, we are interested in the virtualization at the hardware
abstraction layer to emulate a personal computer. Some
well-known commercial PC emulators are KVM1, VMware2,
VirtualBox3 or Virtual PC4.
      </p>
      <p>With this technology it could be possible to virtualize a
group of scaled-down computers using just one real
computer. In this way, with a small cluster of computers (e.g.
20 computers) we could virtualize the whole scaled-down
search engine.</p>
      <p>For example, using a cluster of 20 computers, will
require that each computer virtualizes 50 computers with 10
megabytes of memory and 1 gigabyte of disk. Roughly
speaking, this would take half a gigabyte of memory and
50 gigabytes of disk from the real machine, which should be
easily handled by any modern computer.</p>
      <p>To the best of our knowledge, all the virtualization tools
allow you to set the size of memory and disk for the
virtualized host. Also, some of them (e.g. VMware) can set a limit
for the network usage, in terms of average or peak, and for
the CPU speed.</p>
      <p>From our point of view, these requirements should be
enough to scale-down a host of a distributed search engine.
Some other parameters could also be considered when
scaling down a search engine, such as, disk or memory speed.
We are aware of some solutions in this sense5, but these low
level parameters could be quite hard to virtualize.
However, we doubt about the usefulness of these parameters in
the performance analysis, while the performance of the real
computer is not degraded. In any case, in future works it
1http://www.linux-kvm.org/
2http://www.vmware.com/
3http://www.virtualbox.org/
4http://www.microsoft.com/windows/virtual-pc/
5http://sourceforge.net/apps/trac/ioband/
would be interesting to study in depth the effect of these
low level parameters in the performance analysis of a
scaleddown search engine.</p>
      <p>The use of virtualization is not only interesting to reduce
the resources required to build a scaled-down version of a
search engine. It could be very appealing to test the effect
of new technologies in the performance of a large-scale search
engine. For example, let us say that we want to compare the
performance of the new SSD (Solid State Drive) memories
versus the traditional hard drives. To build a whole search
engine using SSD memories would be a waste of resources
until the performance has been tested. But, buying a cluster
of computers with SSD memory and building a scaled-down
version of the search engine is feasible and not very
expensive. In this way, we could test and compare the performance
of this new technology.
4.</p>
    </sec>
    <sec id="sec-5">
      <title>CONCLUSIONS</title>
      <p>This paper presents a new approach to the performance
evaluation of large-scale search engines based on a
scaleddown version of the distributed system.</p>
      <p>The main problem, when using an analytical or simulation
model for evaluation purposes, is that some (important)
details could be missing in order to make the model feasible
and so, the estimations obtained could differ substantially
from the real values.</p>
      <p>If we use a real search engine for performance evaluation,
the results obtained will be more precise but will depend
on the resources available. A distributed system composed
of a few computers does not constitute a large-scale search
engine and the resources required to build a representative
search engine are excessive for most researchers.</p>
      <p>We suggest to build a scaled-down version of a search
engine using virtualization tools to create a realistic cluster
of computers. By using a scaled-down version of a computer
we expect to maintain the behaviour of the whole distributed
system at the same time that the hardware requirements are
softened. This would be the key to use the virtualization
tools to build a large distributed system using a small cluster
of computers.</p>
      <p>This research is at an early stage, but we strongly believe
that this would be a valid technique to analyse the
performance of a large-scale distributed IR system.</p>
      <p>In the near future we plan to develop a scaled-down search
engine using a small cluster of computers. We would like to
compare the performance of the scaled-down search engine
with an equivalent real search engine to test the accuracy of
this methodology.</p>
    </sec>
    <sec id="sec-6">
      <title>ACKNOWLEDGEMENTS</title>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>C. S.</given-names>
            <surname>Badue</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Baeza-Yates</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Ribeiro-Neto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ziviani</surname>
          </string-name>
          , and
          <string-name>
            <given-names>N.</given-names>
            <surname>Ziviani</surname>
          </string-name>
          .
          <article-title>Analyzing imbalance among homogeneous index servers in a web search system</article-title>
          . Inf. Process. Manage.,
          <volume>43</volume>
          (
          <issue>3</issue>
          ):
          <fpage>592</fpage>
          -
          <lpage>608</lpage>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>R. A.</given-names>
            <surname>Baeza-Yates</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Castillo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Junqueira</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Plachouras</surname>
          </string-name>
          , and
          <string-name>
            <given-names>F.</given-names>
            <surname>Silvestri</surname>
          </string-name>
          .
          <article-title>Challenges on distributed web retrieval</article-title>
          .
          <source>In ICDE</source>
          , pages
          <fpage>6</fpage>
          -
          <lpage>20</lpage>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>F. J. Burkowski.</surname>
          </string-name>
          <article-title>Retrieval performance of a distributed text database utilizing a parallel processor document server</article-title>
          .
          <source>In DPDS '90: Proceedings of the second international symposium on Databases in parallel and distributed systems</source>
          , pages
          <fpage>71</fpage>
          -
          <lpage>79</lpage>
          , New York, NY, USA,
          <year>1990</year>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>F.</given-names>
            <surname>Cacheda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Carneiro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Plachouras</surname>
          </string-name>
          ,
          <string-name>
            <given-names>and I.</given-names>
            <surname>Ounis</surname>
          </string-name>
          .
          <article-title>Performance analysis of distributed information retrieval architectures using an improved network simulation model</article-title>
          .
          <source>Inf</source>
          . Process. Manage.,
          <volume>43</volume>
          (
          <issue>1</issue>
          ):
          <fpage>204</fpage>
          -
          <lpage>224</lpage>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J.-H.</given-names>
            <surname>Chen</surname>
          </string-name>
          and
          <string-name>
            <given-names>C.-C.</given-names>
            <surname>Chang</surname>
          </string-name>
          .
          <article-title>A moving piv system for ship model test in a towing tank</article-title>
          .
          <source>Ocean Engineering</source>
          ,
          <volume>33</volume>
          (
          <fpage>14</fpage>
          -15):
          <fpage>2025</fpage>
          -
          <lpage>2046</lpage>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Chowdhury</surname>
          </string-name>
          and
          <string-name>
            <given-names>G.</given-names>
            <surname>Pass</surname>
          </string-name>
          .
          <article-title>Operational requirements for scalable search systems</article-title>
          .
          <source>In CIKM '03: Proceedings of the twelfth international conference on Information and knowledge management</source>
          , pages
          <fpage>435</fpage>
          -
          <lpage>442</lpage>
          , New York, NY, USA,
          <year>2003</year>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>C. L. A.</given-names>
            <surname>Clarke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Craswell</surname>
          </string-name>
          ,
          <string-name>
            <surname>and I. Soboroff.</surname>
          </string-name>
          <article-title>Overview of the trec 2004 terabyte track</article-title>
          .
          <source>In TREC</source>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J.-H.</given-names>
            <surname>Kang</surname>
          </string-name>
          and
          <string-name>
            <given-names>S.-J.</given-names>
            <surname>Lee</surname>
          </string-name>
          .
          <article-title>Improvement of natural ventilation in a large factory building using a louver ventilator</article-title>
          .
          <source>Building and Environment</source>
          ,
          <volume>43</volume>
          (
          <issue>12</issue>
          ):
          <fpage>2132</fpage>
          -
          <lpage>2141</lpage>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>A.</given-names>
            <surname>Moffat</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Webber</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zobel</surname>
          </string-name>
          , and
          <string-name>
            <given-names>R. A.</given-names>
            <surname>Baeza-Yates</surname>
          </string-name>
          .
          <article-title>A pipelined architecture for distributed text query evaluation</article-title>
          .
          <source>Inf</source>
          . Retr.,
          <volume>10</volume>
          (
          <issue>3</issue>
          ):
          <fpage>205</fpage>
          -
          <lpage>231</lpage>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>B.</given-names>
            <surname>Ribeiro-Neto</surname>
          </string-name>
          and
          <string-name>
            <surname>R. A. Barbosa.</surname>
          </string-name>
          <article-title>Query performance for tightly coupled distributed digital libraries</article-title>
          .
          <source>In Proceedings of the third ACM conference on Digital libraries</source>
          , pages
          <fpage>182</fpage>
          -
          <lpage>190</lpage>
          , Pittsburgh, Pennsylvania, United States,
          <year>1998</year>
          . ACM Press.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>N.</given-names>
            <surname>Susanta</surname>
          </string-name>
          and
          <string-name>
            <given-names>C.</given-names>
            <surname>Tzi-Cker</surname>
          </string-name>
          .
          <article-title>A survey on virtualization technologies</article-title>
          .
          <source>Technical report.</source>
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>A.</given-names>
            <surname>Tomasic</surname>
          </string-name>
          and
          <string-name>
            <given-names>H.</given-names>
            <surname>Garcia-Molina</surname>
          </string-name>
          .
          <article-title>Query processing and inverted indices in shared: nothing text document information retrieval systems</article-title>
          .
          <source>The VLDB Journal</source>
          ,
          <volume>2</volume>
          (
          <issue>3</issue>
          ):
          <fpage>243</fpage>
          -
          <lpage>276</lpage>
          ,
          <year>1993</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>S.</given-names>
            <surname>Zhong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Jabbal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Garcillan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Wood</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C.</given-names>
            <surname>Warsop</surname>
          </string-name>
          .
          <article-title>Towards the design of synthetic-jet actuators for full-scale flight conditions</article-title>
          .
          <source>Flow, Turbulence and Combustion</source>
          ,
          <volume>78</volume>
          (
          <issue>3</issue>
          ):
          <fpage>283</fpage>
          -
          <lpage>307</lpage>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>