<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Resource Oriented Architecture to Handle Data Volume Diversity</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Pierre De Vettor</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Michael Mrissa</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Djamal Benslimane</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Universite de Lyon</institution>
          ,
          <addr-line>CNRS LIRIS, UMR5205, F-69622, France Lyon</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Providing quality-aware techniques for reusing data available on the Web is a major concern for today's organizations. High quality data that o ers higher added-value to the stakeholders is called smart data. Smart data can be obtained by combining data coming from diverse data sources on the Web such as Web APIs, SPARQL endpoints, Web pages and so on. Generating smart data involves complex data processing tasks, typically realized manually or in a static way in current organizations, with the help of statically con gured work ows. In addition, despite the recent advances in this eld, transfering large amounts of data to be processed still remains a tedious task due to unreliable transfer conditions or transfer rate/latency problems. In this paper, we propose an adaptive architecture to generate smart data, and focus on a solution to handle volume diversity during data processing. Our approach aims at maintaining good response time performance upon user request. It relies on the use of RESTful resources and remote code execution over temporary data storage where business data is cached. Each resource involved in data processing accesses the storage to process data on-site.</p>
      </abstract>
      <kwd-group>
        <kwd>resource oriented architecture</kwd>
        <kwd>data integration</kwd>
        <kwd>data semantics</kwd>
        <kwd>smart data</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        During the last few years, governments, companies and organizations have opened
their databases and information systems to the world across the Web, thanks to
initiatives such as the open data project [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. These data sources are typically
exposed via Web APIs [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] or SPARQL endpoints and can be combined in service
mashups [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] to produce highly valuable services. As an/For example, the sets of
APIs provided by Twitter, Amazon, Youtube or Flickr are reused/used again in
thousands of mashups1.
      </p>
      <p>
        This smart use of data has caught the interest of the community as a natural
development. The objective of smart data [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] is focused on producing
highquality data that is directly useful to users. Automatically integrate data from
1 See also http://www.programmableweb.com/
diverse sources in order to produce smart data is currently a hot research topic.
Despite these advances, data quantity still hampers data exchanges and remains
a bottleneck in architectures, especially when it comes to data transfer between
resources. By taking advantage of REST architectural style and the use of
resources, we unfortunately su er from the limitations of Web protocols, and it is
sometimes di cult to transfer large quantities of data through HTTP. There is a
need for an approach that minimizes data transfer and performs data processing
tasks closer to the data source to reduce network tra c.
      </p>
      <p>In this paper, we present our adaptive architecture and propose a solution,
through the use of adaptive data access strategies and remote code execution on
temporary data storage unit resources, to handle latency and architecture issues
that appear when processing data volume diversity. Our architecture optimizes
and adapts work ows to handle the variety of data sources involved in the query.
In this approach, we present this resource oriented architecture and explain our
solution for reducing latency in resource oriented architecture.</p>
      <p>This paper is organized as follows. Section 2 presents related approaches
to handle volume diversity in data sources in RESTful architectures. Section 3
presents our resource oriented architecture and our solution to minimize data
exchange. Section 4 gives an evaluation of our prototype in terms of responsiveness
and shows how it responds to user requests with acceptable timings. Section 5
discusses our results and provides guidelines for future work.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>
        Over the past few years, big data has generated a lot of interest from researchers
and industrials, answering to four challenges, known as the four Vs: Volume,
Variety, Velocity and Veracity. De ning our smart data architecture [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], we provide
data source models and strategies to support the Variety challenge and
analysis, combination and data cleaning for Veracity. Finally, Velocity is managed by
adapting work ows and optimizing resource orchestration at runtime to improve
response time. On top of that, we put more focus on data quality through the
use of semantics, metadata and intelligent processing techniques. In this smart
data context [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], propositions focus on data quality rather than quantity and
build smarter architecture. In this data-driven context, Web standards comes as
a solution, but limits data exchanges. The following appproaches have addressed
the data issue in HTTP-based solutions.
      </p>
      <p>
        Devresse et Al. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] propose a library called libdavix allowing high performance
computing world to bene t from HTTP and the RESTful principles. This
approach focus on adapting the HTTP protocol, maximizing the reutilization of
TCP connections, by providing a dynamic connection pool coupled with the
usage of the HTTP Keep-Alive feature. By avoiding useless protocol handshakes,
reconnections and redirections, their approach improves e ciency of large data
transport through HTTP. In Fast Optimised REST (FOREST) [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], Ko et Al.
propose a non-invasive technique relying on TCP data encapsulation in
UDPbased data transfer payloads [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Evaluations shows good results, but the
approach does not seem to provide a real solution, it is a low level xing to bene t
from advantages of other protocols. Zheng et al. [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] provide an overview of
service-generated big data as well as big data-as-a-service, a exible
infrastructure providing common big data management functionalities. Their approach
rely on cloud computing techniques to handle collection, storing, processing and
visualization of data and they address some signi cant challenges, particularly
about variety or volume and how infrastructure must support (and adapts) this
variety and volume to provide fast data access and processing. Van Der Pol et
Al. [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] propose an approach based on multiple paths TCP [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] to transfer huge
data sets over networks. Their approach relies on load balancing transfer through
the di erent available paths relying on parallel multiple TCP requests. Their
approach can handle di erent paths with di erent bandwidths, balanced over the
di erent interface o ered by the system. They propose a prototype According to
these approaches, it becomes clear that the most powerful solution is to minimize
data transfers, process data volumes closer from the source to reduce the
unnecessary tra c. In the next section, we present our resource-oriented architecture,
the models that helps to build it, and nally, we present our solution to handle
data volume diversity in our smart data architecture.
3
      </p>
    </sec>
    <sec id="sec-3">
      <title>Contribution</title>
      <p>
        In order to handle our smart data challenges, we envision a resource-oriented
architecture, generating adaptive work ows at runtime to adapt to data source
characteristics. We rely on the data source models presented in our previous
work [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] to represent data source characteristics such as uri, request format,
volume, latency, etc. Relying on these models, we are thus able to generate
adaptive data processing work ows according to characteristics that appears on
data source descriptions.
      </p>
      <p>Then we de ne an resource oriented architecture to manage these adaptive
work ows and the resources involved.
3.1</p>
      <sec id="sec-3-1">
        <title>Architecture</title>
        <p>
          In our approach, we de ne di erent necessary steps to complete the data
aggregation process and produce smart data. The main steps are extraction from
data source and transformation into a pivot format, semantic annotation of
the extracted data set ([
          <xref ref-type="bibr" rid="ref2">2</xref>
          ], [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]), combination [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] of obtained data sets,
ltering of data sets in order to remove inconsistent or duplicated data. We divide
each of these tasks as RESTful resources in our architecture and orchestrate
them as work ows. In order to handle data ows between resources, we de ne
an orchestrator, which acts as a data bus. This orchestrator forwards messages
and data from and to resources, builds HTTP request and handles requests
responses. Fig. 1 shows the di erent resources and components of our architecture.
On top of that, we provide our architecture with management API and resources
to avoid manual con gurations as much as possible.
4
In order to handle volume diversity in our RESTful architecture, and to maintain
a good response time performance, we propose a solution which reduces data
transfer to a required minimum.
        </p>
        <p>In our architecture, upon query request, an adapted work ow is generated,
involving resources and services, to handle the required processing tasks. This
work ow is executed, our data bus handles data exchanges, generating HTTP
requests and retrieving responses from resources. We identi ed the following data
transfers during the process: extraction from data sources, data transfer between
core resources, and data download by the user. In this context, data extraction
and download are required, but transfer between resources can be avoided or
modi ed.</p>
        <p>Based on this observation, we modify our architecture to decrease the volume
of exchange data between resources. We design our API to only manipulate
queries and metadata (data models, etc.). Computing this metadata, remote
processing codes are generated in order to complete the process managed by this
resource. These processes are handled close to data, minimizing execution time
by lowering latency and network time.
3.3</p>
      </sec>
      <sec id="sec-3-2">
        <title>Storing data</title>
        <p>In order to temporarily store data, our proposal is based on temporary data
storage units, generated at each user request. Storage units act as le hosting
services, they are generated at runtime, for each new user request, for a limited
amount of time and contains the data and metadata for this request. They are
provided with an engine capable of processing remote processing tasks generated
by resources. Storage units are erased after a certain amount time, which has
been xed to a default value of 24 hours. This delay is customizable for each
request. Time counter is reset each time a user reissues the query or reaccesses
the storage. Storage units are accessible as RESTful resources, for management
purpose or to retrieve query responses when data processing is over. When the
processing tasks are over, the storage URI is given to the user, so that he could
download the data sets answering to his request.
3.4</p>
      </sec>
      <sec id="sec-3-3">
        <title>Processing engine</title>
        <p>In order to handle the di erent tasks required to complete the smart data process,
we provide the storages with a functional engine capable of executing remote
processing tasks directly above the data instances. These codes are generated by
the tasks resources and transfered instead of data.</p>
        <p>
          We rely on functional languages to de ne processing tasks, since our data sets
are stored as tabular data sets (lists or dictionary in JSON or JSON-LD[
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]).
We provide data store with di erent processing engines, each representing an
environment or an engine. Each data store is provided with an API, which
allows its management, but also to register engine libraries or plugins, required
to process the di erent langages or functionnalities This data store is provided
with an API, allowing to register di erent engine libraries or plugin to handle
speci c languages or functionnalities.
4
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Implementation and Evaluation</title>
      <p>In this paper, we focus our work on handling data volume diversity in our
architecture for smart data management.</p>
      <p>In order to evaluate the scalability of our architecture, we demonstrate the
evolution of performance, evaluating request response time when answering to a
set of complex semantic queries over multiple data sources. We vary the number
of data sources and measure response times.
4.1</p>
      <sec id="sec-4-1">
        <title>Use Case Scenario</title>
        <p>We focus our work on the enrichment and reusability of data, and based our
models and implementation on the data handled by a communication company,
which has a need for an adaptive system able to automatically re ne and combine
data coming from their internal information system and enrich it with help from
open Web data sources. This solution has for objectives to study the impact of
campaign broadcasts over a list of customers and to provide decision support
tools for future broadcasts. This scenario describes the following data sources,
presenting several characteristics, speci c to our scenario:</p>
        <p>Source 1 is a linked service giving access to our company small business
data set. Data that come from this source are subject to privacy constraints.
Source 2 is a SQL endpoint to a database that contains millions of tuples
with a low update frequency. Source 3 is another SQL endpoint to a database
that contains customer daily activities, updated regularly. Source 4 is a RSS
stream that contains user update requests (removal requests from customers,
request form architecture's user). Other sources are represented by a set of
Web sources: FOAF and vcard ontologies that help to annotate data, as well as
a Dbpedia SPARQL endpoint. Relying on the scenario presented above, we create
two request, involving di erent concepts subgraphs. We populate our scenario
with a set of data sources, covering the di erent subgraphs.</p>
        <p>PREFIX a l : &lt;http : // r e s t f u l . a l a b s . i o / concepts /0.1/&gt;
PREFIX xsd : &lt;http : //www. w3 . org /TR/xmlschema 2/&gt;
SELECT ? e m a i l v a l u e ? campaign WHERE f
? e m a i l a a l : e m a i l ;
a l : h a s e m a i l v a l u e ? e m a i l v a l u e ;
a l : b l a c k l i s t s t a t u s ? s t a t u s .
? c l i c a a l : c l i c ;
a l : c l i c e m a i l ? e m a i l ;
a l : c l i c d a t e ? date .</p>
        <p>FILTER ( ? s t a t u s != 1 &amp;&amp; ? date &gt;= " 1 4 1 1 4 7 7 4 5 0 " ^^ xsd : date )</p>
        <sec id="sec-4-1-1">
          <title>Listing 1.1. Query 1.1 involving the di erent concepts</title>
          <p>PREFIX a l : &lt;http : // r e s t f u l . a l a b s . i o / concepts /0.1/&gt;
SELECT ? e m a i l v a l u e WHERE f
? e m a i l a a l : e m a i l ;
a l : h a s e m a i l v a l u e ? e m a i l v a l u e ;
a l : b l a c k l i s t s t a t u s ? s t a t u s .</p>
          <p>FILTER ( ? s t a t u s != 1)
g</p>
        </sec>
        <sec id="sec-4-1-2">
          <title>Listing 1.2. Query 1.2 introducing a user speci c lters</title>
          <p>Query 1.1 involves four subgraphs, implying data sources with di erent
characteristics such as high volume (scenario's big database) and privacy sensitive
information (scenario linked service). This query also presents user speci c
lters. Query 1.2 involves only a few number of concepts, and present less data
manipulations. Tests are performed on a double core 2.3 GHz machine, with 4
GB of RAM. Restful resources and architecture are implemented through PHP
frameworks, Zend and Slim2 and hosted on scenario company servers (Apache).
4.2</p>
        </sec>
      </sec>
      <sec id="sec-4-2">
        <title>Evaluation</title>
        <p>We evaluate our architecture response time to query Q1, and Q2 respectively,
when the number of data sources increases. We compared response time evolution
for the same queries and data, with and without our optimisation process.</p>
        <p>As can be seen in Fig. 2 and Fig. 3, the volume of data increase with the
number of data source, and the classical approach response time su er, due to
HTTP transfers. In parallel, optimized approach response time increase linearly,
but stays acceptable standing under the treshold of 4 seconds. In this case,
transfer and network latency consumes time exponentially. Query Q2 involves
less concepts, and as a result less data manipulations, but our architecture still
su er from a very high latency due to data transfer. When the number of data
source exceed 24, the classical approach is unable to provide a response in an
acceptable delay. This graph clearly show that our architecture can adapt to
large volumes of data, using our optimisation technique.
2 See http://www.slimframework.com/
) 10
s
(
e
m
i
.t 5
p
s
e
R
0
8
)
(s 6
e
it 4
m
.
p
se 2
R
0</p>
        <sec id="sec-4-2-1">
          <title>4 sources</title>
        </sec>
        <sec id="sec-4-2-2">
          <title>Classic</title>
        </sec>
        <sec id="sec-4-2-3">
          <title>8 sources 12 sources</title>
        </sec>
        <sec id="sec-4-2-4">
          <title>Optimized</title>
          <p>16 sources 20 sources
24 sources
In this paper, we describe a resource-oriented architecture that relies on
adaptive data processing strategies to optimize data exchanges between resources that
process data. We focus on the latency problems that appear when dealing with
diverse data volumes especially when transferring data between the di erent
architecture components. All along the smart data construction process, we rely on
a temporary data storage where business data is stored. Our architecture
handles the communication between the di erent resources, by transferring metadata
about the query and the data storage URI. Resources generate adapted remote
processing codes, which are forwarded to storage units and executed on-site by
the data storage engine. Therefore, data manipulation is performed on-site, and
the data does not ow through the architecture. By reducing latency due to
data transfers, we alleviate our decentralized and distributed architecture from
the burden of data tranfer, and improve the responsiveness of data-driven
resource work ows. We support our implementation with a set of evaluations and
tests through our use case scenario. As future work, we envision to investigate
the appearance of data uncertainty in data aggregating approach, relying on
probabilistic integration techniques.</p>
        </sec>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <given-names>Djamal</given-names>
            <surname>Benslimane</surname>
          </string-name>
          , Schahram Dustdar, and
          <string-name>
            <given-names>Amit P.</given-names>
            <surname>Sheth</surname>
          </string-name>
          .
          <article-title>Services mashups: The new generation of web applications</article-title>
          .
          <source>IEEE Internet Computing</source>
          ,
          <volume>12</volume>
          (
          <issue>5</issue>
          ):
          <volume>13</volume>
          {
          <fpage>15</fpage>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <given-names>Christian</given-names>
            <surname>Bizer</surname>
          </string-name>
          .
          <article-title>D2rq - treating non-rdf databases as virtual rdf graphs</article-title>
          .
          <source>In In Proceedings of the 3rd International Semantic Web Conference (ISWC2004</source>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3. Pierre De Vettor, Michael Mrissa, and
          <string-name>
            <given-names>Djamal</given-names>
            <surname>Benslimane</surname>
          </string-name>
          .
          <article-title>Models and architecture for smart data management</article-title>
          .
          <source>Technical Report RR-LIRIS-2014-017, LIRIS UMR 5205 CNRS/Universite Claude Bernard Lyon 1/Universite Lumiere Lyon</source>
          <volume>2</volume>
          /Ecole Centrale de Lyon,
          <year>December 2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <given-names>Adrien</given-names>
            <surname>Devresse</surname>
          </string-name>
          and
          <string-name>
            <given-names>Fabrizio</given-names>
            <surname>Furano</surname>
          </string-name>
          .
          <article-title>E cient HTTP based I/O on very large datasets for high performance computing with the libdavix library</article-title>
          .
          <source>CoRR, abs/1410.4168</source>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <given-names>A</given-names>
            <surname>Ford</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C</given-names>
            <surname>Raiciu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M</given-names>
            <surname>Handley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S</given-names>
            <surname>Barre</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J</given-names>
            <surname>Iyengar</surname>
          </string-name>
          .
          <article-title>Architectural guidelines for multipath tcp development</article-title>
          .
          <source>In IETF, RFC 6182. Citeseer</source>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Robert L. Grossman</surname>
          </string-name>
          , Yunhong Gu, Xinwei Hong, Antony Antony, Johan Blom, Freek Dijkstra, and Cees de Laat.
          <article-title>Tera ows over gigabit fWANsg with fUDTg</article-title>
          .
          <source>Future Generation Computer Systems</source>
          ,
          <volume>21</volume>
          (
          <issue>4</issue>
          ):
          <volume>501</volume>
          {
          <fpage>513</fpage>
          ,
          <year>2005</year>
          .
          <article-title>High-Speed Networks</article-title>
          and
          <article-title>Services for Data-Intensive Grids: the DataTAG Project</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7. Lushan Han,
          <string-name>
            <given-names>Tim</given-names>
            <surname>Finin</surname>
          </string-name>
          , Cynthia Parr, Joel Sachs, and Anupam Joshi.
          <article-title>Rdf123: From spreadsheets to rdf</article-title>
          .
          <source>In The Semantic Web - ISWC</source>
          <year>2008</year>
          , volume
          <volume>5318</volume>
          of Lecture Notes in Computer Science, pages
          <volume>451</volume>
          {
          <fpage>466</fpage>
          . Springer Berlin Heidelberg,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <given-names>Tom</given-names>
            <surname>Heath</surname>
          </string-name>
          and
          <string-name>
            <given-names>Christian</given-names>
            <surname>Bizer</surname>
          </string-name>
          .
          <article-title>Linked Data: Evolving the Web into a Global Data Space</article-title>
          .
          <article-title>Synthesis Lectures on the Semantic Web</article-title>
          . Morgan &amp; Claypool Publishers,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>R.K.L. Ko</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Kirchberg</surname>
          </string-name>
          ,
          <string-name>
            <surname>Bu-Sung Lee</surname>
            ,
            <given-names>and E.</given-names>
          </string-name>
          <string-name>
            <surname>Chew</surname>
          </string-name>
          .
          <article-title>Overcoming large data transfer bottlenecks in restful service orchestrations</article-title>
          .
          <source>In Web Services (ICWS)</source>
          ,
          <year>2012</year>
          IEEE 19th International Conference on, pages
          <volume>654</volume>
          {
          <fpage>656</fpage>
          ,
          <year>June 2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <given-names>Markus</given-names>
            <surname>Lanthaler</surname>
          </string-name>
          and
          <string-name>
            <given-names>Christian</given-names>
            <surname>Gutl</surname>
          </string-name>
          .
          <article-title>On using json-ld to create evolvable restful services</article-title>
          .
          <source>In Proceedings of the Third International Workshop on RESTful Design</source>
          , WS-REST '
          <volume>12</volume>
          , pages
          <fpage>25</fpage>
          {
          <fpage>32</fpage>
          , New York, NY, USA,
          <year>2012</year>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>M. Maleshkova</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Pedrinaci</surname>
            , and
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Domingue</surname>
          </string-name>
          .
          <article-title>Investigating web apis on the world wide web</article-title>
          .
          <source>In Web Services (ECOWS)</source>
          ,
          <source>2010 IEEE 8th European Conference on</source>
          , pages
          <volume>107</volume>
          {
          <fpage>114</fpage>
          ,
          <string-name>
            <surname>Dec</surname>
          </string-name>
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Michael</surname>
            <given-names>Mrissa</given-names>
          </string-name>
          , Mohamed Sellami, Pierre De Vettor, Djamal Benslimane, and
          <string-name>
            <given-names>Bruno</given-names>
            <surname>Defude</surname>
          </string-name>
          .
          <article-title>A decentralized mediation-as-a-service architecture for service composition</article-title>
          .
          <source>2012 IEEE 21st International Workshop on Enabling Technologies: Infrastructure for Collaborative Enterprises</source>
          ,
          <volume>0</volume>
          :
          <fpage>80</fpage>
          {
          <fpage>85</fpage>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <given-names>Amit</given-names>
            <surname>Sheth</surname>
          </string-name>
          .
          <article-title>Transforming big data into smart data: Deriving value via harnessing volume, variety, and velocity using semantic techniques and technologies</article-title>
          .
          <source>In Data Engineering (ICDE)</source>
          ,
          <source>2014 IEEE 30th International Conference on, pages 2{2</source>
          ,
          <string-name>
            <surname>March</surname>
          </string-name>
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14. R. van der Pol,
          <string-name>
            <given-names>S.</given-names>
            <surname>Boele</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Dijkstra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Barczyk</surname>
          </string-name>
          ,
          <string-name>
            <surname>G. van Malenstein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. H.</given-names>
            <surname>Chen</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.</given-names>
            <surname>Mambretti</surname>
          </string-name>
          .
          <article-title>Multipathing with mptcp and open ow. High Performance Computing, Networking Storage and Analysis</article-title>
          ,
          <source>SC Companion:</source>
          ,
          <volume>0</volume>
          :
          <fpage>1617</fpage>
          {
          <fpage>1624</fpage>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Zibin</surname>
            <given-names>Zheng</given-names>
          </string-name>
          , Jieming Zhu, and
          <string-name>
            <given-names>M.R.</given-names>
            <surname>Lyu</surname>
          </string-name>
          .
          <article-title>Service-generated big data and big data-as-a-service: An overview</article-title>
          .
          <source>In Big Data (BigData Congress)</source>
          ,
          <source>2013 IEEE International Congress on</source>
          , pages
          <volume>403</volume>
          {
          <fpage>410</fpage>
          ,
          <year>June 2013</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>