<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>DataPACT: Compliance by Design of Data/AI Operations and Pipelines⋆</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Dumitru Roman</string-name>
          <email>dumitru.roman@sintef.no</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>George Konstantinidis</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Matteo Palmonari</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Marta Musidlowska</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Radu Prodan</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>KU Leuven Centre for IT &amp; IP Law</institution>
          ,
          <addr-line>Sint-Michielsstraat 6 box 3443, 3000 Leuven</addr-line>
          ,
          <country country="BE">Belgium</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>SINTEF AS</institution>
          ,
          <addr-line>Forskningsveien 1, 0373 Oslo</addr-line>
          ,
          <country country="NO">Norway</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>University of Innsbruck</institution>
          ,
          <addr-line>Technikerstraße 21a, 9020 Innsbruck</addr-line>
          ,
          <country country="AT">Austria</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>University of Southampton.</institution>
          <addr-line>Highfield, SO17 1BJ, Southampton</addr-line>
          ,
          <country country="UK">United Kingdom</country>
        </aff>
      </contrib-group>
      <fpage>39</fpage>
      <lpage>47</lpage>
      <abstract>
        <p>DataPACT is a key initiative that develops novel tools and methodologies for efficient, compliant, ethical, and sustainable data/AI operations and pipelines. DataPACT contributes to their design, implementation, and management by embedding compliance, privacy, and environmental sustainability at their core design. It delivers compliance-by-design for data/AI operations and pipelines by developing innovative technical tools (Compliance Toolbox) and supportive methodologies (Compliance Framework) for compliance assessment and realization of data/AI pipelines designed, deployed, and executed through a set of management tools and techniques (Compliance-aware Data/AI Pipeline Toolbox). This paper presents an overview of DataPACT, focusing on motivation, methodology, and use cases.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Data/AI operations</kwd>
        <kwd>pipelines</kwd>
        <kwd>compliance</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        In an era when ubiquitously generated and extensively utilized data drives decision-making and
innovation, ensuring the compliance, fairness, and environmental sustainability of data/AI
operations has become paramount. The disruptive emergence of new AI models, the increasing
volume of data, the complexity and computational needs of AI systems, the interaction of different
and often competing actors, and the multitude of emerging legislations pose significant compliance
challenges with regulations such as the General Data Protection Regulation [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], the Data
Governance Act [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], the Data Act [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], and the Artificial Intelligence Act [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Moreover, there is an
increasing societal and business demand for ethical and transparent AI and an urgent need to
mitigate the environmental impact of data and AI operations aligning with the European Green
Deal’s objectives [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>Traditionally, data/AI pipeline developments neglect inherent mechanisms to ensure compliance
with various legislations and ethical guidelines and commonly treat compliance as an afterthought.
This oversight presents a significant challenge as the demand for compliance in data operations
grows alongside the increasing complexity and scale of data and AI systems. Integrating
compliance, fairness, and environmental sustainability into data/AI operations is more critical than
ever, especially considering the European Green Deal’s ambitious goals. Addressing the gap
between existing data/AI pipeline architectures and the requirements for legal, ethical, and
environmental compliance remains an unsolved problem, underscoring the urgency and relevance
of initiatives such as DataPACT. Recent developments highlight the complexity of retrofitting
compliance into data/AI pipelines and the necessity for new, inherently compliant frameworks.
The increased complexity associated with including AI operations significantly affects the vast
numbers of small to large organizations willing to exploit AI’s potential but may be overwhelmed
by compliance regulations.</p>
      <p>DataPACT is a new initiative funded through the Horizon Europe program that aims to address
these challenges by developing and piloting novel tools and methodologies that enable efficient,
compliant, ethical, and environmentally friendly data/AI operations and pipelines. In this context:
(1) Data/AI operations involve comprehensive processes and methodologies to discover, collect,
integrate, process, store, manage, and analyze data for deriving valuable insights and powering
artificial intelligence (AI) systems (such operations encompass a wide range of activities, including
data discovery, cleaning, linking, model training, and deployment of AI models); (2) Data/AI
pipelines refer to the automated sequences of data processing and AI model training steps designed
to efficiently manage data flow from its raw form to a usable state for analytics, machine learning,
and AI applications (pipelines ensure scalable, repeatable, and efficient data processing, analysis,
and utilization, crucial for AI systems’ systematic handling and functioning).</p>
      <p>DataPACT envisions an approach where compliance, ethics, and environmental sustainability
are not afterthoughts but foundational elements of data/AI operations and pipelines. This
transformation ensures a future where companies and public sector organizations can effortlessly
ensure compliance with legal and ethical standards and unlock the full potential of their data
assets, fostering an environment of trust and transparency for citizens. DataPACT is a holistic
approach to addressing the multifaceted challenges of compliance, trust, and environmental
sustainability in the data-driven world, paving the way for a more ethical, transparent, and
environmentally conscious digital future. DataPACT develops innovative technical tools and
toolsupported methodologies for compliance assessment and realization of data/AI pipelines designed,
deployed, and executed through a set of pipeline management tools and techniques
(Complianceaware Data/AI Pipeline Toolbox).</p>
      <p>This paper discusses provides an easy-to-understand motivating example (Section 2), outlines
the DataPACT approach (Section 3), discusses relevant use cases (Section 4), and provides a
summary and outlook (Section 5).</p>
    </sec>
    <sec id="sec-2">
      <title>2. DataPACT motivating example</title>
    </sec>
    <sec id="sec-3">
      <title>3. DataPACT methodology</title>
      <p>DataPACT develops methods to support the compliance needs of data and AI operations and
pipelines. It relies on primitive, compliance-aware data and AI operations and an infrastructure for
combining them in executable analysis data/AI pipelines that achieve business goals and create
value for organizations. Primitive operations and pipelines interact by design with socio-technical
tools for compliance assessment and realization. Operation APIs introduced as wrappers on top of
existing operations support the execution and transition to compliance-ensuring data/AI pipelines.
DataPACT thus offers capabilities for the design, management, execution, and monitoring of
data/AI pipelines that consider relevant compliance aspects. The DataPACT pipeline lifecycle
management toolbox consists of conceptual design and runtime phases where algorithms or
methods can either assess or enforce compliance aspects of data and AI operations or entire
pipelines, e.g., privacy-preservation, explainability, consent management, trustworthiness legality,
fairness, transparency, energy-efficiency, adherence to (smart) contracts, and others. Assessment
tools are an integral part of the methodology to support the interpretation of the pipeline and its
result or certification of compliance by reusing and creating a machine-processable vocabulary for
different legislations and guidelines.</p>
      <p>DataPACT core elements. DataPACT operates with three elements: data, data/AI operations,
and data/AI pipelines combined to realize business goals. DataPACT offers tools and techniques for
data, operations, and pipelines to enable their management and operations compliant with GDPR,
AI Act, Data Act, and Data Governance Act. Big Data in DataPACT (primarily volume, variety, and
velocity) can be highly unstructured or structured. Examples of data operations include data
discovery, profiling, access, cleaning, linking, validation, and structuring. Examples of AI
operations include algorithm selection, feature engineering, hyperparameter tuning, and model
training. Data/AI pipelines are combinations of such operations applied to data.</p>
      <p>CEURART uses the Libertinus fonts. You may have to install these fonts on your computer. The
text below shows how to locally install them.</p>
      <sec id="sec-3-1">
        <title>3.1. Conceptual architecture</title>
        <p>Typical data/AI pipelines managed by DataPACT involve a feedback sequence of steps separated
by design (Identify, define, assess, realize, deploy) and runtime (execute, monitor, reassess, realize)
involving a set of relevant stakeholders, depicted in Figure 2.
1. Identify data/AI operations: To perform business operations, business domain experts use
domain-specific knowledge to identify relevant data and conduct operations. Data/AI
scientists with AI and ML expertise specify implementation details of the operations, such
as analytical models and operation-specific code. Special-purpose visual tools can add
compliance-related annotations to the datasets and operations.
2. Define data/AI pipelines: During pipeline definition, business domain experts and data/AI
scientists use the data processing requirements to structure, configure, design, and simulate
pipelines, assisted by AI tools, like LLM-based generation from high-level textual
specifications.
3. Assess compliance of data/AI operations/pipelines: After designing the pipelines, compliance,
data protection, ethics, and risk officers analyze the pipelines in the light of the triggered
regulations and respective guidelines (GDPR, AI Act, Data Act) through the legal and
ethical assessment framework. LLM-assisted tools extract relevant information from the
target regulations and guidelines, recommending actions to make the pipeline compliant.
4. Realize data/AI operations/pipeline compliance: Compliance engineers get recommendations
for the assessment process from the previous step and implement compliance
recommendations in pipeline design, possibly refining or changing them through the
DataPACT tools.
5. Deploy data/AI pipelines: After ensuring pipeline compliance in the previous step, the
data/AI engineers will provide the hardware/software infrastructure for automatically
deploying and executing the pipeline using the DataPACT tools.
6. Execute data/AI operations: After deploying the pipeline, its execution starts with the input
data from the selected data providers, depicted with the dashed arrows at the bottom of the
figure. DataPACT provides the necessary tools for ensuring scalable and secure execution
of the pipeline.
7. Monitor data/AI pipelines: Monitoring the pipeline during the execution triggers compliance
reassessment if the execution results in data or steps are uncompliant.
8. Reassess compliance of data/AI operations/pipelines: In case of an uncompliant pipeline at
run time, the Compliance, Data Protection, Ethics, and Risk Officers initiate a compliance
assessment process similar to step (3) analyzing the execution of the pipeline against the
target regulations, guides and issuing guidelines for ensuring compliance at runtime.
9. Realize compliance of data/AI operations/pipelines: Similar to step 4, compliance engineers
implement the step 8 recommendations by changing or adapting pipeline execution. This
step ensures that execution results (data and insights) comply with target regulations and
guidelines in data consumers’ interest.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. DataPACT toolbox, framework and sandbox</title>
        <p>DataPACT supports these through technical tools and methodologies for compliance assessment
and realization of data/AI pipelines, designed, deployed, and executed and through a set of pipeline
management tools and techniques:
•
•
•</p>
        <p>Compliance toolbox and framework cover several technical tools relevant to compliance
(e.g., privacy policies, consent management, negotiation, bias) and methodologies
supporting legislation, ethics, and social impact.</p>
        <p>Compliance-aware data/AI pipeline toolbox supports several tools to manage the lifecycle of
data/AI pipelines (e.g., identification of relevant operations, design, deployment).
Stakeholders use such tools to build generic pipelines for flagship applications, such as
using and fine-tuning foundation models or sharing data in data spaces.</p>
        <p>Regulatory sandbox environments test the envisioned toolboxes and framework under
regulatory supervision and ensure appropriate safeguards before the use case deployment
(e.g., healthcare, customer support). DataPACT validates the technology-supporting
compliance of data/AI pipelines for eight use cases from various domains, detailed in the
next section.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Artifacts and tools</title>
        <p>
          Development of compliance-aware data/AI pipelines envisions:
1. Relevant, tested, and documented data/AI operations, embedding compliance-related
metadata into data/AI operations and pipelines;
2. Tool for compliance-aware data/AI pipeline design with visual specification interfaces and
novel LLM-assisted generation from high-level textual descriptions and requirements;
3. Tools for compliance-aware simulation, deployment, and secure execution of data/AI
operations/pipelines (based on [
          <xref ref-type="bibr" rid="ref7 ref8">7,8</xref>
          ]);
4. Tool for monitoring and traceability of data, operations, and pipelines.
        </p>
        <p>
          Privacy, consent, and access policies for data/AI pipelines envisions:
1. Fine-grained rule-based language and policy enforcement tool for specifying and enforcing
access and use policies (e.g., privacy, legislation), recognizing violations in data/AI
operations/pipelines, and recommending policy-compliant amendments;
2. Consent management tool for collecting, storing, and managing (e.g., updating,
withdrawing) consent regarding personally identifiable data that dynamically adapts the
consent-compliant execution of data/AI operations/pipelines;
3. Machine-processable contract language and tool for managing, automatically negotiating
and enforcing algorithmic contracts/agreements containing detailed properties of data/AI
operations and pipelines (e.g., fine-grained statements, price, energy consumption).
Trust, fairness, robustness, and explainability for compliance in data/AI pipelines envision:
1. Tool for managing trust and reputation scores of stakeholders involved in data/AI pipelines
for data sharing/processing contracts and agreements;
2. Tool for the interactive inspection of biases in the data based on statistical analysis
3. Tool for the declarative specification of fairness constraints for traditional ML approaches
based on a neuro-symbolic approach (based on [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]);
4. Tool for simplified assessment of fairness, robustness, and quality of LLMs, guarding
against misbehaviors of generative LLMs;
5. Tool for explainability of data/AI operations and pipelines powered by LLMs.
Compliance-aware environmental sustainability of data/AI operations/pipelines envisions:
1. Tool for reporting energy consumption at different granularities of data/AI operations and
pipelines and recommendation engine for suggesting energy improvements;
2. Framework for trade-off analysis between performance and energy efficiency in AI training
and inference pipelines;
3. Framework for trade-off analysis between energy efficiency and
privacy/fairness/robustness of data/AI pipelines.
        </p>
        <p>Legal, ethical, and social impact assessment envisions:
1. Framework for legal assessment of data, operations, and pipelines based on mapping and
analysis of legal provisions that trigger specific compliance requirements;
2. Framework for ethical and responsibility assessment of data, operations, and pipelines,
based on assessment methodologies for human agency, diversity, societal well-being, and
respect for individual rights, emphasizing non-discrimination and proper redress;
3. Sector- and use case-based methodology for social impact assessment of data/AI pipelines,
engaging diverse and inclusive perspectives, involving different stakeholders, and utilizing
interdisciplinary approaches to comprehend the specific impact of data/AI pipelines;
4. LLM-assisted tool for understanding compliance of data, operations, and pipelines based on
natural language descriptions of legislations, regulations, ethical guidelines, and social
impact;
5. Tool for legal, ethical, and environmental compliance certification of data, operations,
pipelines, and associated stakeholders.</p>
        <p>Demonstrate the usefulness in flagship complaint applications of data/AI pipelines envisions:
1. LLMOps pipeline tool for fine-tuning foundation models with built-in compliance use;
2. RAG-LLM pipeline for retrieval augmented generation improving the accuracy of
contextualized foundation models (avoiding hallucinations);
3. Pipeline for supporting data-sharing workflows in data spaces.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. DataPACT use cases</title>
      <p>DataPACT initially selected eight use cases (UCs) to develop and validate the proposed approach,
toolboxes, and framework, whereby it targeted a balanced mix of commercial (UC1, UC5, UC6,
UC7) and public sector (UC2, UC3, UC4, UC8) organizations and data, and a high degree of
coverage of regulations and their triggers, with high coverage of at least two principal regulations
for each UC. DataPACT has selected UCs, which, due to their complexity in terms of the variety of
technologies used and their innovative nature, will serve as illustrative examples to analyze the
application of the various provisions of the new regulatory framework in practice and to test tools
to facilitate compliance with these requirements.</p>
      <p>Table 1 summarizes a preliminary analysis of the relevance of the selected UCs, focusing on the
regulations that set the legal framework for the data economy and development of data/AI
pipelines. In that context, “triggers” are contextual elements that lead to applying the legal
framework when present or performing a specific function, as relevant in the context of
DataPACT. Triggers indicated in the table impact the overall development and operation of the
respective UCs or specific aspects of AI/data pipelines within the UCs, which require meeting
requirements set by the regulation.
•
•
•
•
•
•
•
•</p>
      <p>Media and entertainment estimate the expected impact (number of viewers and
visualizations) of new media content (movies, series, advertising) using brain-computer
interface technologies optimized in the cloud-edge continuum for different use scenarios.
Healthcare develops a new AI decision support system to predict high-risk adverse health
outcomes of patients after hospitalization and the optimal day of discharge, providing
healthcare providers essential information on patients’ functional status when
transitioning to home care.</p>
      <p>Smart city develops a new service for compliance assessment of data pipelines, delivering
data in the Urban Data Space and AI pipelines consuming data from the data space and
contributing to the compliance assessment of connectors as the key component of the data
space.</p>
      <p>Law enforcement and security develop a new AI solution to reduce the time and effort to
analyze datasets with personal data, sensitive information, and specific watermarks for
classified data and automatically determine the security classification level based on
internal/national policies.</p>
      <p>Customer relationship develops an AI-based call center for customer relationship
management, integrating voice and voice-to-text with customer data, marketing data, and
financial metrics, performing call NLP and sentiment analysis to enhance customer support
and business strategies.</p>
      <p>Manufacturing, develops a new generative AI service that improves service delivery
productivity for medical imaging system devices, fined-tuned with specific knowledge and
employed in a trustworthy, transparent, cost-efficient manner.</p>
      <p>Human resources develop a chatbot for employees and HR managers enquiring about
personal and operative employee information, including an AI system for decision-making
processes, such as promotions, monitoring and performance evaluation, and work-related
relationships.</p>
      <p>Public data designs processes and pipelines for making currently restricted datasets owned
and managed by municipalities available to support academic and industrial research and
share data with public organizations and the private sector.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Summary and outlook</title>
      <p>This paper argued the need for novel tools and methodologies that enable efficient, compliant,
ethical, and sustainable data/AI operations and pipelines. We introduced DataPACT as a key
initiative in this context and provided relevant challenges, a motivating example, the proposed
approach, and example use cases. With a broad target group including Data/AI industries, Data/AI
scientists, business experts, MLOps/DataOps, policymakers, law practitioners, compliance and data
protection officers, DataPACT addresses the lack of dedicated software tools and methodologies for
supporting complaint data/AI operations and pipelines required to build compliant data-driven
AIenabled applications. DataPACT aims to deliver robust and practical software toolboxes and
frameworks to support complaint data/AI pipelines, including a compliance-aware data pipelines
toolbox for making data/AI pipelines compliance-aware, a compliance toolbox for auto-mating
compliance-related tasks, and a compliance framework for le-gal/ethical/social assessment,
validated in proper use cases products and services: media, movie making, healthcare, AI-assisted
customer service, data spaces, banking, medical device services, law enforcement and security,
public administration.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgements</title>
      <p>This work was partly funded through the Horizon Europe projects DataPACT (101189771),
enRichMyData (101070284), UPCAST (101093216), and Austrian Research Promotion Agency “AIM
AT Stiftungsprofessur für Edge AI” (909989). The authors acknowledge the contributions of all
DataPACT project partners in developing the project.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used ChatGPT in order to partly: Grammar
and spelling check, Paraphrase and reword. After using this tool/service, the author(s)
reviewed and edited the content as needed and take(s) full responsibility for the publication’s
content.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>General</given-names>
            <surname>Data Protection Regulation</surname>
          </string-name>
          , Regulation - 2016/679 - EN
          <string-name>
            <surname>- GDPR -</surname>
          </string-name>
          EUR-Lex, https://eur-lex.europa.eu/eli/reg/2016/679/oj (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Data</given-names>
            <surname>Governance</surname>
          </string-name>
          <string-name>
            <given-names>Act</given-names>
            , Regulation
            <surname>-</surname>
            2022/868 - EN -
          </string-name>
          EUR-Lex, https://eur-lex.europa.eu/eli/reg/2022/868/oj/eng (
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Data</given-names>
            <surname>Act</surname>
          </string-name>
          , Regulation - EU -
          <year>2023</year>
          /2854 - EN
          <string-name>
            <surname>-</surname>
          </string-name>
          EUR-Lex, https://eur-lex.europa.eu/eli/reg/2023/2854/oj/eng (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Artificial</given-names>
            <surname>Intelligence</surname>
          </string-name>
          <string-name>
            <surname>Act</surname>
          </string-name>
          , Regulation - EU -
          <year>2024</year>
          /1689 - EN
          <string-name>
            <surname>-</surname>
          </string-name>
          EUR-Lex, https://eurlex.europa.eu/eli/reg/2024/1689/oj/eng (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>The</given-names>
            <surname>European Green Deal</surname>
          </string-name>
          ,
          <string-name>
            <surname>EUR-Lex -</surname>
          </string-name>
          52019DC0640
          <string-name>
            <surname>- EN -</surname>
          </string-name>
          EUR-Lex, https://eur-lex.europa.eu/legal-content/EN/TXT/?uri=CELEX:52019DC0640 (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>N.</given-names>
            <surname>Nikolov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Solberg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Prodan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Soylu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Matskin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Roman</surname>
          </string-name>
          ,
          <article-title>Container-based data pipelines on the computing continuum for remote patient monitoring</article-title>
          .
          <source>in Computer</source>
          , vol.
          <volume>56</volume>
          , no.
          <issue>10</issue>
          , pp.
          <fpage>40</fpage>
          -
          <lpage>48</lpage>
          , Oct.
          <year>2023</year>
          , doi: 10.1109/
          <string-name>
            <surname>MC</surname>
          </string-name>
          .
          <year>2023</year>
          .
          <volume>3285414</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>D.</given-names>
            <surname>Roman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Prodan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Nikolov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Soylu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mataskin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Marella</surname>
          </string-name>
          ,
          <article-title>Big Data Pipelines on the Computing Continuum: Tapping the Dark Data,"</article-title>
          <source>in Computer</source>
          , vol.
          <volume>55</volume>
          , no.
          <issue>11</issue>
          , pp.
          <fpage>74</fpage>
          -
          <lpage>84</lpage>
          , Nov.
          <year>2022</year>
          , doi: 10.1109/
          <string-name>
            <surname>MC</surname>
          </string-name>
          .
          <year>2022</year>
          .
          <volume>3154148</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>A.</given-names>
            <surname>Thomas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Nikolov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Pultier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Roman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Elvesaeter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Soylu</surname>
          </string-name>
          ,
          <string-name>
            <surname>SIM-PIPE</surname>
            <given-names>DryRunner</given-names>
          </string-name>
          :
          <article-title>An approach for testing container-based big data pipelines and generating simulation data"</article-title>
          .
          <source>2022 IEEE 46th Annual Computers, Software, and Applications Conference (COMPSAC)</source>
          , Los Alamitos, CA, USA,
          <year>2022</year>
          , pp.
          <fpage>1159</fpage>
          -
          <lpage>1164</lpage>
          , doi: 10.1109/COMPSAC54236.
          <year>2022</year>
          .
          <volume>00182</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>G.</given-names>
            <surname>Greco</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Palmonari</surname>
          </string-name>
          ,
          <article-title>Declarative Encoding of Fairness in Logic Tensor Networks</article-title>
          .
          <source>26th European Conference on Artificial Intelligence</source>
          ,
          <string-name>
            <surname>ECAI</surname>
          </string-name>
          <year>2023</year>
          . IOS Press,
          <year>2023</year>
          , pp.
          <fpage>908</fpage>
          -
          <lpage>915</lpage>
          . doi:
          <volume>10</volume>
          .3233/FAIA230360.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>