<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>THE DESIGNING OF CLOUD INFRASTRUCTURE CONSISTING OF GEOGRAPHICALLY DISTRIBUTED DATA CENTERS</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>P.V. Fedchenkov</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>S.E. Khoruzhnikov</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>N.Y. Samokhin</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>A.Y. Shevel</string-name>
          <email>shevel_a_y@niuitmo.ru</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Network and Cloud Technologies, ITMO University</institution>
          ,
          <addr-line>St.-Petersburg, 197101</addr-line>
          ,
          <country country="RU">Russia</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>National Research Centre "Kurchatov Institute" PETERSBURG NUCLEAR PHYSICS INSTITUTE</institution>
          ,
          <addr-line>Gatchina, 188300</addr-line>
          ,
          <country country="RU">Russia</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>2018 Petr V. Fedchenkov, Sergey E. Khoruzhnikov</institution>
          ,
          <addr-line>Nikita Y. Samokhin</addr-line>
          ,
          <country>Andrey Y. Shevel</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2018</year>
      </pub-date>
      <fpage>32</fpage>
      <lpage>36</lpage>
      <abstract>
        <p>Contemporary tendency of use the cloud architectures is quite common. Many tests show that the price for using a commercial cloud is feasible for most of applications including High Energy Physics. At the same time, many users are still using the large data centers (DC). The using just one data center (DC) is not completely sufficient in several aspects. First is lack of isolation, and second is lack of elasticity. Lack of isolation means situation when many users share one DC and violate SLA of each other from time to time. Lack of elasticity means situation when many users while sharing one DC cannot change SLA at required level. In context of service availability, one DC for many users is also not perfect solution as technical, nature, and social incidents might affect the DC. As a result, cloud solution based on geographically distributed DCs looks preferable. The running project addressed to implementation of such the cloud is described in the paper.</p>
      </abstract>
      <kwd-group>
        <kwd>clouds</kwd>
        <kwd>SDN</kwd>
        <kwd>NFV</kwd>
        <kwd>distributed</kwd>
        <kwd>QKD</kwd>
        <kwd>software defined</kwd>
        <kwd>linux</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>The number of DCs in the world is steadily growing. There are many attempts in many areas
to move applications from concrete DC to public cloud. Using multiple geographically distributed
DCs looks preferable from reliability and resource availability points of view. The network connection
between DCs included into the cloud should correspond with a SLA for the cloud: if a data transfer
speed is serious priority, the data links might have a bandwidth of 100 Gbit or more. In order to
implement the cloud, the special management system is required.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Main architecture approaches and technical requirements</title>
      <p>
        The number of modern approaches: Software Defined Storage (SDS), Software defined
Network (SDN), Network Function Virtualization (NFV), Infrastructure as Code (IaC) [
        <xref ref-type="bibr" rid="ref1 ref2">1,2</xref>
        ] are
proved in many developments. Technical requirements are necessary for describing the future system
features as well. Among technical requirements for system management of Scalable Geographically
Distributed DCs (SGDD), the following is considered as most important:
 Data links
o
o
o
o
o
      </p>
      <sec id="sec-2-1">
        <title>Hardware security.</title>
      </sec>
      <sec id="sec-2-2">
        <title>Data compression/decompression.</title>
      </sec>
      <sec id="sec-2-3">
        <title>Data encryption/decryption.</title>
      </sec>
      <sec id="sec-2-4">
        <title>Service reliability</title>
      </sec>
      <sec id="sec-2-5">
        <title>Steady system functioning.</title>
        <p>Gradual degradation with hardware or software malfunction.</p>
      </sec>
      <sec id="sec-2-6">
        <title>Monitoring of physical and virtual infrastructure.</title>
        <p>Data storage features, including the long-term data storage with total volume up to 1 EB.</p>
      </sec>
      <sec id="sec-2-7">
        <title>SLA support.</title>
      </sec>
      <sec id="sec-2-8">
        <title>Data Center Infrastructure Management (DCIM).</title>
      </sec>
      <sec id="sec-2-9">
        <title>Automatic deployment of the system management</title>
        <p>System Management Architecture of SGDD is implemented in form of separate program
agents communicating with each other by the special designed protocol. System logical structure is
shown in fig. 1. In the paper, the technical term program agent or agent means program daemon
which is running in isolated operating environment. Any request to the agent and answer form the
agent could be performed with appropriate protocol. Such implementation gives many advantages:
 independence for agent development: developer can use any programming language
and/or library or versions of it;
independence during system production: if one agent is down, other agents might
continue to work, i.e. system is just partially degraded;
simplification of horizontal scalability: it is possible to set up an additional agent
instance with same functions in order to improve service bandwidth and to provide
high availability and load balancing.</p>
        <p>The system functioning principles are shown in fig. 1 as well. The user (client) can be signed
up at the portal of registration and the accounting agent. After registration user is able to create a
subscription where required virtual facilities are described. Such facilities may include virtual
machines, virtual storages, virtual data links. All facilities may be accompanied by specific SLAs.
Initially, all requests are headed to orchestration agent which does preliminary check of available
resources. After that, user subscription request is routed to appropriate service agent: VM allocation
agent, virtual data links allocation agent, storage allocation agent. Those service agents perform the
creation of required virtual objects and send back the technical information how to access created
virtual objects to the user.</p>
        <p>For example, if creation of virtual storage with specific SLA is required, the storage allocation
agent does create the CEPH cluster with required SLA including special gateway, which allows the
user to read and write the data to and from virtual storage.</p>
        <p>
          Virtual data links allocation agent does create new virtual links with specific SLA. SLA for
virtual links may include encryption/decryption, compression/decompression, etc. The implementation
of virtual data links is shown in fig. 2. The agent issues some commands for RYU SDN controller [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]
which at the same time does control an Open Virtual Switch (OVS) for choosing an appropriate
network segment. All network segments must be created before the system is in operation. Practically
data links creation is performed in form of the choice of appropriate network segment.
        </p>
      </sec>
      <sec id="sec-2-10">
        <title>Virtual objects of type VM are created in similar ways.</title>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Technical solutions for system management of SGDD</title>
      <p>


</p>
      <p>Main technical solutions within technical designing are:
The system is based on Free and Open Source Software (FOSS) components.</p>
      <p>
        Operating System: Naulinux (compatible with RedHat, Scientific Linux, CentOS) [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
Computing, portals: Openstack [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>
        Data Storage - Software Defined Storage (SDS): CEPH [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].









      </p>
      <p>
        Monitoring of physical and virtual infrastructure, billing data: Zabbix [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>
        Visualization: Grafana [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]/Kibana [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
      </p>
      <p>Database keeping the hardware information, virtual objects information, detailed requests and
answers information: PostgreSQL.</p>
      <p>Communication within SGDD: RabbitMQ with special protocol developed.</p>
      <p>
        Data transfer security: symmetric cryptography with OpenSSL and Quantum Key Distribution
(QKD) procedure [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
      </p>
      <p>Independence of external components changes (versions, polices, etc): own versions of Linux
distribution (Naulinux) + code repository for all developed components.</p>
      <p>Each type of service must be feasible for accounting/billing purposes.</p>
      <p>Current debugging of the system is performed providing the following hardware test
configuration:
 Several micro DCs consisting of servers [6 + 2 + 2] type HPE DL380 Gen10 8LFF CTO (each
server has 2x Intel Zeon Gold 6130, main memory: 128 GB, HD storage 32 TB).
Several switches: HP FF 5700-32XGT-8XG-2QSFP+ Switch.</p>
      <p>Other communication equipment.</p>
      <p>
        Automated deployment of the system has been implemented with SALTstack [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] and special
scripts developed.
      </p>
    </sec>
    <sec id="sec-4">
      <title>4. Conclusion</title>
      <p>Main features of the project are: the use of FOSS components, system architecture based on
the independent agents, QKD procedure, own code repository, automated deployment
of the system [11]. An engineering infrastructure monitoring might permit to use data centers as dark
data centers. Here dark data center means absence of permanent maintaining staff on the data center. It
may be some technical persons are visiting such the data center to fix something by demand. All
features together provide more security and reliability. The project is in active debugging stage now.
The system is being prepared for a preproduction stage by first half of the 2019.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Acknowledgement</title>
      <p>The development is performed in accordance with the contract № 03.G25.31.0229
“Development of new technological components for management systems of geographically
distributed Data Centers”
[11] P.V. Fedchenkov et al // APPROACHES TO THE AUTOMATED DEPLOYMENT OF THE
CLOUD INFRASTRUCTURE OF GEOGRAPHICALLY DISTRIBUTED DATA CENTERS // these
proceedings.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Matej</given-names>
            <surname>Artac</surname>
          </string-name>
          et al //
          <article-title>Infrastructure-as-Code for Data-Intensive Architectures: A Model-Driven Development Approach // 2018</article-title>
          <source>IEEE International Conference on Software Architecture (ICSA) 30 April-4 May</source>
          <year>2018</year>
          // DOI: 10.1109/ICSA.
          <year>2018</year>
          .
          <volume>00025</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Kief</given-names>
            <surname>Morris</surname>
          </string-name>
          // Infrastructure as Code: Managing Servers in the Cloud // ISBN-13:
          <fpage>978</fpage>
          -
          <lpage>1491924358</lpage>
          , ISBN 9781491924358 // O'Reilly Media //
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Naulinux</surname>
          </string-name>
          . Available at: http://naulinux.ru
          <source>/ (accessed 01 November</source>
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4] OpenStack. Available at: http://openstack.org
          <source>/ (accessed 01 November</source>
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Wong M.-T.</surname>
          </string-name>
          et al //
          <article-title>Ceph as WAN Filesystem - Performance and Feasibility Study through Simulation // Network Research WorkshopProceedings of the Asia-Pacific Advanced Network 2014 v</article-title>
          .
          <volume>38</volume>
          , p.
          <fpage>1</fpage>
          -
          <lpage>11</lpage>
          . http://dx.doi.
          <source>org/10.7125/APAN.38.1 ISSN 2227-3026.</source>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Zabbix</surname>
          </string-name>
          . Available at: http://zabbix.org/wiki/Main_Page (accessed
          <issue>01</issue>
          <year>November 2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Grafana</surname>
          </string-name>
          . Available at: https://grafana.com
          <source>/ (accessed 01 November</source>
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Kibana</surname>
          </string-name>
          . Available at: https://www.elastic.co/products/kibana
          <source>(accessed 01 November</source>
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>G. P.</given-names>
            <surname>Miroshnichenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. V.</given-names>
            <surname>Kozubov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. A.</given-names>
            <surname>Gaidash</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. V.</given-names>
            <surname>Gleim</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D. B.</given-names>
            <surname>Horoshko</surname>
          </string-name>
          , “
          <article-title>Security of subcarrier wave quantum key distribution against the collective beam-splitting attack”</article-title>
          <source>Opt. Express</source>
          <volume>26</volume>
          ,
          <fpage>11292</fpage>
          -
          <lpage>11308</lpage>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>SaltStack</surname>
          </string-name>
          . Available at: https://saltstack.com/community/ (accessed
          <issue>01</issue>
          <year>November 2018</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>