<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>FAIR Data Based on Extensible Unifying Data Model Development</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>Proceedings of the XX International Conference “Data Analytics and Management in Data Intensive Domains” (DAMDID/RCDL'2018)</institution>
          ,
          <addr-line>Moscow</addr-line>
          ,
          <country country="RU">Russia</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Sergey Stupnikov © Leonid Kalinichenko Institute of Informatics Problems, Federal Research Center “Computer Science and Control“ of the Russian Academy of Sciences</institution>
          ,
          <addr-line>Moscow</addr-line>
          ,
          <country country="RU">Russia</country>
        </aff>
      </contrib-group>
      <fpage>9</fpage>
      <lpage>13</lpage>
      <abstract>
        <p>Nowadays data sources within data infrastructures are quite heterogeneous, they are represented using very different data models. Data models vary from relational one to NoSQL zoo of data models. A prerequisite for (meta)data interoperability, integration and reuse within some data infrastructure is unification of source data models and their data manipulation languages. A unifying data model (called canonical) has to be chosen for the data infrastructure. Every source data model has to be mapped into the canonical model, mapping should be formalized and verified. The paper overviews data unification techniques have been developed during recent years and discusses application of these techniques for data integration within FAIR data infrastructures.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1 Introduction</title>
      <p>Data sources nowadays are quite heterogeneous: they
are represented using very different data models. Variety
of data models includes traditional relational model and
its object-relational extensions, array and graph-based
models, semantic models like RDF and OWL, models for
semi-structured data like NoSQL, XML, JSON and so
on. These models provide also very different data
manipulation and query languages for accessing and
modifying data.</p>
      <p>A prerequisite for (meta)data interoperability,
integration and reuse within some data infrastructure is
unification of source data models and their data
manipulation languages. A unifying data model (called
canonical) has to be chosen for the data infrastructure.
The canonical data model serves as the language for
knowledge representation mentioned in FAIR I1
principle ((meta)data use a formal, accessible, shared,
and broadly applicable language for knowledge
representation) [1][2]. Every source data model has to be
mapped into the canonical model. Mapping can be
accompanied with the extension of the canonical model
if required. A mapping should be formalized and
verified: a formal proof that the mapping preserves
semantics of data structures and data manipulation
operations of the source data model should be provided.</p>
      <p>As the core of the canonical model some concrete
data model like SQL (conforming to ISO/ANSI SQL
standard of 2011 or later) or RDF/RDF Schema with
SPARQL query language can be used. To cover features
of various source data models the canonical model has to
be extensible. Examples of extensions are specific data
structures (data types), compound operations or
restrictions (dependencies). An extension is constructed
for every source data model. Canonical model is formed
as the union of the core data model and all extensions.</p>
      <p>
        Data unification techniques were extensively studied
at FRC CSC RAS [3]. As the core of the canonical model
specific object-frame language with broad range of
modeling facilities was used [4]. Approaches for
mapping of different classes of source data models were
developed: process models [5], semantic models [
        <xref ref-type="bibr" rid="ref10">6</xref>
        ][
        <xref ref-type="bibr" rid="ref17">13</xref>
        ],
array [
        <xref ref-type="bibr" rid="ref13">9</xref>
        ] and graph-based [
        <xref ref-type="bibr" rid="ref14 ref2">10</xref>
        ] models, some other kinds
of NoSQL models [
        <xref ref-type="bibr" rid="ref12">8</xref>
        ]. Techniques for verification of
mappings applying a formal language based on the first
order logic and set theory and supported by automatic
and interactive provers were developed [
        <xref ref-type="bibr" rid="ref15">11</xref>
        ][
        <xref ref-type="bibr" rid="ref16">12</xref>
        ].
      </p>
      <p>Techniques mentioned are proposed as a formal basis
for (meta)data interoperability, integration and reuse
within FAIR data infrastructures. Such infrastructures
may combine virtual integration facilities (subject
mediators) as well as data warehouses to integrate
heterogeneous data sources in an interoperable way.</p>
      <p>The rest part of the paper is structured as follows:
section 2 overviews data unification techniques that have
been developed during recent years and section 3
discusses application of these techniques for data
integration within FAIR data infrastructures.</p>
    </sec>
    <sec id="sec-2">
      <title>2 Data Model Unification</title>
      <p>Various source data models and their data manipulation
languages applied within some data infrastructure have
to be unified in the frame of some canonical data model.</p>
      <p>
        The main principle of the canonical model design
(synthesis) for a data infrastructure is the extensibility of
the canonical model kernel in heterogeneous
environment [3], including various models used for the
representation of resources of the data infrastructure. A
kernel of the canonical model is fixed (for instance, SQL
or RDF). A specific source data model R of the
environment is said to be unified if it is mapped into the
canonical model C [
        <xref ref-type="bibr" rid="ref15">11</xref>
        ][
        <xref ref-type="bibr" rid="ref16">12</xref>
        ]. This means a creation of
such extension E of the canonical model kernel (note that
such extension can be empty) and such mapping M of a
source model into extended canonical one that the source
model refines the extended canonical one. Model
refinement of C by R means that for any admissible
specification (schema) r represented in R its image M(r)
in C under the mapping M is refined by the specification
r. Such refining mapping of models means preserving of
operations and information of a source model while
mapping it into the canonical one. Preserving of
operations and information should be formally proven.
The canonical model for the environment is synthesized
as the union of extensions, constructed for all models of
the environment.
      </p>
      <p>The following languages and formal methods are
required to support data model mapping:
• a kernel of the canonical data model;
• formal methods allowing to describe data model
syntax as well as semantic mappings
(transformations) of one model to another;
• formal methods supporting verification of refinement
reached by the mapping.</p>
      <p>Within studies on data unification techniques at FRC
CSC RAS as a kernel of the canonical data model the
SYNTHESIS language [4] was used. The SYNTHESIS
language, as a hybrid semistructured and object-oriented
data model, includes the following distinguishing
features: facilities for definitions of frames, abstract data
types, classes and metaclasses, functions and processes,
logical formulae facilities applied for description of
constraints, queries, pre- and post-conditions of
functions, assertions related to processes. For extension
of the canonical model kernel, metaclasses, metaframes,
parameterized constructions including assertions and
generic data types were applied. Data unification
teqhniques developed can be adopted also for other
canonical data model kernels like SQL or RDF.</p>
      <p>
        For data model’s semantics formalization and
refinement verification the AMN (Abstract Machine
Notation) language [
        <xref ref-type="bibr" rid="ref18">14</xref>
        ] was used. The language is
supported by technology and tools for proving of
refinement (B-technology) [
        <xref ref-type="bibr" rid="ref19">15</xref>
        ]. AMN is based on the
first order predicate logic and Zermelo-Frenkel set
theory and enables to consider state space specifications
and behavior specifications in an integrated way. The
system state is specified by means of state variables and
invariants over these variables, system behavior is
specified by means of operations defined as generalized
substitutions – a sort of predicate transformers.
Refinement of AMN specifications is formalized as a set
of refinement proof obligations – theorems of first order
logic. Generally speaking in terms of pre- and
postconditions of operations, refinement of AMN
specifications means weakening pre-conditions and
strengthening post-conditions of corresponding
operations included in these specifications. Proof
requests are generated automatically and should be
proven with the help automatic and interactive theorem
prover [
        <xref ref-type="bibr" rid="ref19">15</xref>
        ].
      </p>
      <p>For the formal description of model syntax and
transformations two approaches were developed and
prototyped.</p>
      <p>
        The first approach [
        <xref ref-type="bibr" rid="ref15">11</xref>
        ][
        <xref ref-type="bibr" rid="ref16">12</xref>
        ] is based on the
metacompilation languages SDF (Syntax Definition
Formalism) and ASF (Algebraic Specification
Formalism). For the languages a tool support —
MetaEnvironment [
        <xref ref-type="bibr" rid="ref20">16</xref>
        ] — is provided based on term rewriting
techniques. Data model syntax is represented using SDF
in a version of extended Backus–Naur form. Data model
transformations are defined as ASF language modules
constituted by sets of functions. A function defines a
transformation of a syntactic element of a source model
into a syntactic element of the canonical model.
Recursive calls of transformation functions are allowed.
According to the ASF-definition the transformation
program code (C language) is generated automatically by
means of Meta-Environment tools. The transformation
obtained is used for mapping of source model
specifications into the canonical model specifications.
      </p>
      <p>
        The second approach [
        <xref ref-type="bibr" rid="ref21">17</xref>
        ] is based on the
ModelDriven Architecture (MDA) [
        <xref ref-type="bibr" rid="ref22">18</xref>
        ] proposed by Object
Management Group. Data model abstract syntax
neglecting any syntactic sugar is defined using Ecore
metamodel (an implementation of OMG's Essential
Meta-Object Facility) used in Eclipse Modeling
Framework [
        <xref ref-type="bibr" rid="ref23">19</xref>
        ]. Concrete syntax of data models binding
syntactic sugar and abstract syntax was for-malized
using EMFText framework [
        <xref ref-type="bibr" rid="ref24">20</xref>
        ]. Data model
transformations are defined using ATLAS
Transformation Language (ATL) [
        <xref ref-type="bibr" rid="ref25">21</xref>
        ] combining
declarative and imperative features. ATL transformation
programs are composed of rules that define how source
model elements are matched and navigated to create and
initialize the elements of the target models. Type system
of the ATL is very close to the type system of the OMG
Object Constraint Language.
      </p>
      <p>Using both approaches construction of a mapping of
a source data model R into the canonical model C is
divided into the following stages:
• formalization of syntax and semantics the models R
and C (if the latter has not yet been defined);
• definition of reference schemas of the models R and</p>
      <p>C (if the latter has not yet been defined);
• integration of reference schemas of the model R and</p>
      <p>C;
• creation of a required extension E of the canonical
model C;
• construction of a transformation of the model R into
the extended canonical model;
• verification of refinement of the extended canonical
model by the model R.</p>
      <p>The Reference schema of a data model is an abstract
description containing concepts related to constructs of
the model and significant associations among these
concepts. Both concepts and associations may be
annotated by verbal definitions (looking like entries in an
explanatory dictionary). Using MDA terms reference
schemas are just metamodels conforming the Ecore
metamodel.</p>
      <p>Formalization of data model semantics and
verification of data model refinement can be performed
in two ways.</p>
      <p>
        In the first way formalization of data model semantics
means a construction of transformations of source and
canonical data model specifications into
AMNspecifications. So for any specification of a source data
model the AMN-specification expressing its semantics is
generated automatically (for instance, in [
        <xref ref-type="bibr" rid="ref15">11</xref>
        ][
        <xref ref-type="bibr" rid="ref16">12</xref>
        ] the
Ontology Web Language [
        <xref ref-type="bibr" rid="ref26">22</xref>
        ] is considered as a source
model and its semantics in AMN is illustrated by
example). Also, for any specification of the canonical
data model the AMN-specification expressing its
semantics is generated automatically [
        <xref ref-type="bibr" rid="ref27">23</xref>
        ]. After that
refinement of a canonical data model specification by a
source data model specification is reduced to refinement
of their semantic AMN specifications and can be verified
by the refinement theorem prover [
        <xref ref-type="bibr" rid="ref19">15</xref>
        ]. So verification of
model refinement is realized over a set of source model
specification samples.
      </p>
      <p>
        In the second way semantics of a data model (source
or canonical) as a whole is expressed by an AMN
specification. For instance, in [
        <xref ref-type="bibr" rid="ref13">9</xref>
        ] AMN semantics for an
array data model is defined, in [
        <xref ref-type="bibr" rid="ref14 ref2">10</xref>
        ] AMN semantics for
a graph data model is defined. AMN semantics for the
SYNTHESIS language as the canonical data model was
also provided [
        <xref ref-type="bibr" rid="ref13">9</xref>
        ][
        <xref ref-type="bibr" rid="ref14 ref2">10</xref>
        ]. Data structures used in data
models were represented by variables in AMN
specifications, properties of data structures were
represented by AMN invariants, typical operations of
data models were represented by AMN operations.
Generally refinement of the AMN-specification MC
corresponding to the canonical data model C by the
AMN-specification MR corresponding to a source data
model R should be also proved using refinement theorem
prover [
        <xref ref-type="bibr" rid="ref19">15</xref>
        ].
      </p>
      <p>
        Partial automation of data unification techniques
mentioned above was implemented within Unifying
Information Models Constructor (Model Unifier in short)
[
        <xref ref-type="bibr" rid="ref15">11</xref>
        ][
        <xref ref-type="bibr" rid="ref16">12</xref>
        ]. Unifier consists of the following main
components:
• tool for the formal description and correctness
checking of model syntax and transformations
(MetaEnvironment, ATL Tools);
• Atelier B [
        <xref ref-type="bibr" rid="ref19">15</xref>
        ], supporting AMN and providing
facilities for proving of specification refinement;
• model manager.
      </p>
      <p>Meta-Environment, ATL Tools and Atelier B are
third-party products. Model manager provides a
graphical interface allowing an expert to search for, view
and register data models and extensions of the canonical
model; to call specific components for generating
templates, editing and integration of reference schemas,
generating templates for translators of source models
into the canonical one, translation of source models
specifications into AMN or into canonical specifications,
translation of canonical specifications into AMN.</p>
      <p>Recent years data unification techniques were
applied to wide range of source data models. In [5] a
canonical process model has been synthesized for the
environment of workflow patterns classified by W. M. P.
van der Aalst. Thus the canonical process model
possesses a property of completeness with respect to
broad class of process models used in various Workflow
Management Systems as well as the languages used for
process composition of Web services.</p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref15">11</xref>
        ][
        <xref ref-type="bibr" rid="ref16">12</xref>
        ] the Ontology Web Language was unified
with the SYNTHESIS language, in [
        <xref ref-type="bibr" rid="ref10">6</xref>
        ] OWL 2 QL was
mapped into the SYNTHESIS.
      </p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref11">7</xref>
        ] application of the canonical model synthesis
methods for the value inventive data models was
discussed. The distinguishing feature of these data
models is inference of new, unknown values in the
process of query answering.
      </p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref12">8</xref>
        ] an approach to mapping of different types of
NoSQL models into the object model of the
SYNTHESIS language used as unifying data model was
considered.
      </p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref13">9</xref>
        ] unification of an array-based data model used
in SciDB DBMS was considered, and in [
        <xref ref-type="bibr" rid="ref14 ref2">10</xref>
        ] unification
of an attributed graph data model was considered. For
both models verification using AMN specifications is
provided.
      </p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref17">13</xref>
        ] issues on unification of RDF with
accompanying RDF Schema and SPARQL languages
were discussed.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3 FAIR Data Based on Data Model</title>
    </sec>
    <sec id="sec-4">
      <title>Unification</title>
      <p>The following levels of integration (from higher to
lower) can be distinguished: data model integration
(unification), schema matching and integration
(metadata integration) and data integration proper.
Usually completion of the integration on a higher level is
a prerequisite for integration on a lower level. Obviously
the highest level, i.e. data model unification is a
prerequisite for (meta)data interoperability, integration
and reuse within FAIR data infrastructures and data
model unification techniques overviewed in the previous
section can be considered as a formal basis for achieving
FAIRness of data.</p>
      <p>Any level of integration makes data more FAIR:
integrated data are much easier to find, access and reuse
and also integrated data are more interoperable than
heterogeneous data stored in different data sources. The
most mature level of integration is achieved within data
integration systems like subject mediators or data
warehouses.</p>
      <p>
        Subject mediators implement virtual integration with
user queries defined in some unified data model. Such
queries are to be decomposed into sets of subqueries and
these subqueries are to be transferred to heterogeneous
data sources. Data sources are connected with a subject
mediator via wrappers which transforms queries into
source data models and also transforms query answers
from source data models into unified mediator data
model. Query answers are transferred by wrappers back
to the mediator, combined and sent to users. One of the
latest trends nowadays is construction of subject
mediators over data lakes [
        <xref ref-type="bibr" rid="ref28">24</xref>
        ].
      </p>
      <p>Data warehouses implement materialized integration
with all required data extracted from sources,
transformed into unified warehouse data model, and
stored into a warehouse.</p>
      <p>Any kind of integration system requires unified data
model. One of the important issues to be resolved for data
integration within FAIR data infrastructures is the choice
of the canonical model kernel. Even the choice between
SQL and RDF is difficult. On the one hand SQL is
supported by industrial standards, methods and
technologies evolving for decades. On the other hand,
RDF is W3C Recommendation supported by triplestore
vendors, is strongly connected with OWL ontological
framework, allows flexible evolution of data schema,
provides logic inference in a native way that is very
important for knowledge bases.</p>
      <p>
        To integrate heterogeneous data sources in an
interoperable way FAIR data infrastructures may support
both mentioned kinds of data integration systems and
also combined data integration systems [
        <xref ref-type="bibr" rid="ref29">25</xref>
        ] with data
warehouses considered as resources to be integrated
within subject mediators. For all kinds of data integration
systems the data model unification techniques can
provide a formal basis.
      </p>
      <p>Acknowledgments. The research is partially supported
by Russian Foundation for Basic Research, project
1807-01434.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <source>stewardship. Sci. Data</source>
          <volume>3</volume>
          :
          <fpage>160018</fpage>
          (
          <year>2016</year>
          ). DOI:
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          10.1038/sdata.
          <year>2016</year>
          .
          <volume>18</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <surname>PeerJ Preprints</surname>
          </string-name>
          (
          <year>2016</year>
          ). URL:
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>https://doi.org/10.7287/peerj.preprints.2522v1</mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <surname>University</surname>
          </string-name>
          ,
          <year>2004</year>
          . -- P.
          <fpage>101</fpage>
          -
          <lpage>116</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <given-names>Resource</given-names>
            <surname>Environments</surname>
          </string-name>
          . Moscow: IPI RAN,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>- 171 p.</mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <source>LNCS 3631</source>
          , pp.
          <fpage>183</fpage>
          -
          <lpage>198</lpage>
          . Berlin-Heidelberg:
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          Springer-Verlag (
          <year>2005</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Kalinichenko</surname>
            ,
            <given-names>L.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stupnikov</surname>
            ,
            <given-names>S.A.</given-names>
          </string-name>
          :
          <article-title>OWL as Yet Another Data Model to be Integrated</article-title>
          .
          <source>In: Advances in Databases and Information Systems: Proc. II of the 15th East-European Conference</source>
          , pp.
          <fpage>178</fpage>
          -
          <lpage>189</lpage>
          . Vienna: Austrian Computer Society (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Kalinichenko</surname>
            <given-names>L.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stupnikov</surname>
            <given-names>S.A.</given-names>
          </string-name>
          :
          <article-title>Synthesis of the Canonical Models for Database Integration Preserving Semantics of the Value Inventive Data Models</article-title>
          .
          <source>Advances in Databases and Information Systems: Proc. of the 16th East European Conference, LNCS 7503</source>
          , pp.
          <fpage>223</fpage>
          -
          <lpage>239</lpage>
          . BerlinHeidelberg: Springer-Verlag (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Skvortsov</surname>
            <given-names>N. A.</given-names>
          </string-name>
          :
          <article-title>Mapping of NoSQL data models to object specifications</article-title>
          .
          <source>Proc. of the 14th Russian Conference on Digital Libraries RCDL'2012. CEUR Workshop Proceedings</source>
          <volume>934</volume>
          :
          <fpage>53</fpage>
          -
          <lpage>62</lpage>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Stupnikov</surname>
            ,
            <given-names>S. A.</given-names>
          </string-name>
          :
          <article-title>Unification of an Array Data Model for the Integration of Heterogeneous Information Resources</article-title>
          .
          <source>In: Proc. of the 14th Russian Conference on Digital Libraries RCDL'2012. CEUR Workshop Proceedings</source>
          , Vol.
          <volume>934</volume>
          , pp.
          <fpage>42</fpage>
          -
          <lpage>52</lpage>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Stupnikov</surname>
            ,
            <given-names>S. A.</given-names>
          </string-name>
          :
          <article-title>Mapping of a Graph Data Model into an Object-Frame Canonical Information Model for the Development of Heterogeneous Information Resources Integration Systems</article-title>
          .
          <source>In: Proc. of the 15th Russian Conference on Digital Libraries RCDL'2013. CEUR Workshop Proceedings</source>
          <volume>1108</volume>
          :
          <fpage>85</fpage>
          -
          <lpage>94</lpage>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Zakharov</surname>
            ,
            <given-names>V. N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kalinichenko</surname>
            ,
            <given-names>L. A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sokolov</surname>
            ,
            <given-names>I. A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stupnikov</surname>
            ,
            <given-names>S. A.</given-names>
          </string-name>
          :
          <source>Development of Canonical Information Models for Integrated Information Systems. Informatics and Applications</source>
          ,
          <volume>1</volume>
          (
          <issue>2</issue>
          ):
          <fpage>15</fpage>
          -
          <lpage>38</lpage>
          (
          <year>2007</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Kalinichenko</surname>
            ,
            <given-names>L.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stupnikov</surname>
            ,
            <given-names>S.A.</given-names>
          </string-name>
          :
          <article-title>Constructing of Mappings of Heterogeneous Information Models into the Canonical Models of Integrated Information Systems</article-title>
          .
          <source>In: Advances in Databases and Information Systems: Proc. of the 12th EastEuropean Conference</source>
          , pp.
          <fpage>106</fpage>
          -
          <lpage>122</lpage>
          . Pori: Tampere University of Technology (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Skvortsov</surname>
            <given-names>N. A.</given-names>
          </string-name>
          :
          <article-title>Mapping of RDF Data Model into the Canonical Model of Subject Mediators</article-title>
          .
          <source>Proc. of the 15th Russian Conference on Digital Libraries RCDL'2013. CEUR Workshop Proceedings</source>
          <volume>1108</volume>
          :
          <fpage>95</fpage>
          -
          <lpage>101</lpage>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Abrial</surname>
          </string-name>
          ,
          <string-name>
            <surname>J.-R.: The B-Book</surname>
          </string-name>
          : Assigning Programs to Meanings. Cambridge: Cambridge University Press (
          <year>1996</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Atelier</surname>
            <given-names>B,</given-names>
          </string-name>
          <article-title>the industrial tool to efficiently deploy the B Method</article-title>
          . http://www.atelierb.eu/
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Van den Brand M. G. J.</surname>
          </string-name>
          et al.:
          <article-title>The ASF+SDF meta-environment: a component based language development</article-title>
          environment // Compiler Construction 2001 / Ed. By R. Wilhelm, pp.
          <fpage>365</fpage>
          -
          <lpage>370</lpage>
          . Springer (
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Stupnikov</surname>
            ,
            <given-names>S.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kalinichenko</surname>
            ,
            <given-names>L.A.</given-names>
          </string-name>
          :
          <article-title>Methods for Semi-automatic Construction of Information Models Transformations</article-title>
          .
          <source>Proc. of the 13th EastEuropean Conference Advances in Databases and Information Systems</source>
          , workshop Model - Driven
          <source>Architecture: Foundations, Practices and Implications (MDA)</source>
          , pp.
          <fpage>432</fpage>
          -
          <lpage>440</lpage>
          . Riga: Riga Technical University (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>Object</given-names>
            <surname>Management Group Model Driven</surname>
          </string-name>
          <article-title>Architecture (MDA)</article-title>
          .
          <source>MDA Guide rev. 2.0. OMG Document ormsc/2014-06-01</source>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Steinberg</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Budinsky</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Paternostro</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Merks</surname>
          </string-name>
          , E.: EMF:
          <article-title>Eclipse Modeling Framework, 2nd Edition</article-title>
          . Addison-Wesley
          <string-name>
            <surname>Professional</surname>
          </string-name>
          (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>EMFText</given-names>
            <surname>Concrete Syntax</surname>
          </string-name>
          <article-title>Mapper</article-title>
          . https://github.com/DevBoost/EMFText
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [21]
          <article-title>ATL - a model transformation technology</article-title>
          . https://eclipse.org/atl/
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>OWL</given-names>
            <surname>Web Ontology Language Reference. W3C Recommendation</surname>
          </string-name>
          . http://www.w3.org/TR/owl-ref/ (
          <year>2004</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [23]
          <string-name>
            <surname>Stupnikov</surname>
            ,
            <given-names>S. A.</given-names>
          </string-name>
          :
          <article-title>A semantic transformation of the canonical information model into a formal specification langage for the refinement verification</article-title>
          .
          <source>Proc. of the 12th Russian Conference on Digital Libraries RCDL'2010</source>
          , pp.
          <fpage>383</fpage>
          -
          <lpage>391</lpage>
          . Kazan: Kazan Federal University (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [24]
          <string-name>
            <surname>Hai</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Quix</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhou</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Query Rewriting for Heterogeneous Data Lakes</article-title>
          . In: Benczúr A.,
          <string-name>
            <surname>Thalheim</surname>
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Horváth</surname>
            <given-names>T</given-names>
          </string-name>
          . (eds) Advances
          <source>in Databases and Information Systems. ADBIS 2018. LNCS</source>
          <volume>11019</volume>
          :
          <fpage>35</fpage>
          -
          <lpage>49</lpage>
          . Springer (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [25]
          <string-name>
            <surname>Stupnikov</surname>
            ,
            <given-names>S.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vovchenko</surname>
            ,
            <given-names>A.E.</given-names>
          </string-name>
          :
          <article-title>Combined Virtual and Materialized Environment for Integration of Large Heterogeneous Data Collections</article-title>
          .
          <source>In: 16th Russian Conference on Digital Libraries RCDL 2014 Proceedings. CEUR Workshop Proceedings</source>
          <volume>1297</volume>
          :
          <fpage>201</fpage>
          -
          <lpage>210</lpage>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>