<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>International Journal of Interactive Mobile Technologies (iJIM) 17 (2023) pp. 125-139.
doi:10.3991/ijim.v17i23.43479.
[26] P. Singh</journal-title>
      </journal-title-group>
      <issn pub-type="ppub">1000-1239</issn>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="doi">10.23919/SoftCOM50211.2020.9238323</article-id>
      <title-group>
        <article-title>Automating machine learning: A meta-synthesis of MLOps tools, frameworks and architectures</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Danylo O. Hanchuk</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Serhiy O. Semerikov</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Academy of Cognitive and Natural Sciences</institution>
          ,
          <addr-line>54 Universytetskyi Ave., Kryvyi Rih, 50086</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Institute for Digitalisation of Education of the NAES of Ukraine</institution>
          ,
          <addr-line>9 M. Berlynskoho Str., Kyiv, 04060</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Kryvyi Rih National University</institution>
          ,
          <addr-line>11 Vitalii Matusevych Str., Kryvyi Rih, 50027</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Kryvyi Rih State Pedagogical University</institution>
          ,
          <addr-line>54 Universytetskyi Ave., Kryvyi Rih, 50086</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>Zhytomyr Polytechnic State University</institution>
          ,
          <addr-line>103 Chudnivsyka Str., Zhytomyr, 10005</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2020</year>
      </pub-date>
      <volume>12168</volume>
      <fpage>362</fpage>
      <lpage>414</lpage>
      <abstract>
        <p>Automating the end-to-end lifecycle of machine learning models is critical for their efective operationalization. Various tools, frameworks and architectures have emerged to support Machine Learning Operations (MLOps) practices. This paper presents a meta-synthesis of existing reviews to provide a comprehensive overview of such enabling technologies for MLOps. The capabilities and features ofered by common commercial and open-source MLOps platforms are compared. Patterns in the MLOps architecture and design philosophies are identified. The role of containers, orchestration, configuration management, and infrastructure automation in ML pipelines is examined. Approaches for model deployment on cloud and edge are also discussed. The synthesis ofers insights for tool selection and usage to automate enterprise-scale machine learning.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;MLOps</kwd>
        <kwd>automation</kwd>
        <kwd>tools</kwd>
        <kwd>frameworks</kwd>
        <kwd>architecture</kwd>
        <kwd>model deployment</kwd>
        <kwd>ML pipelines</kwd>
        <kwd>meta-synthesis</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        In the modern world, machine learning is becoming an increasingly important technology that finds
application in various fields such as finance, healthcare, industry, retail, etc. [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] However, despite
significant progress in the development of machine learning algorithms and models, their efective
deployment in production environments remains a challenging task [
        <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
        ]. This is due to a number of
factors, such as the need to ensure scalability, reproducibility, security, and reliability of models, as well
as the complexity of integrating development and operation processes.
      </p>
      <p>
        To solve these problems, the MLOps (Machine Learning Operations) methodology has emerged, which
aims to apply the principles and practices of DevOps to the development and deployment processes
of machine learning models [
        <xref ref-type="bibr" rid="ref4 ref5">4, 5</xref>
        ]. MLOps covers a wide range of practices, such as automation of
machine learning pipelines, versioning of data and models, monitoring model performance, experiment
management, etc. [
        <xref ref-type="bibr" rid="ref6 ref7">6, 7</xref>
        ]. Research shows that the application of MLOps practices can significantly
increase the eficiency and reliability of deploying machine learning models in production environments
[
        <xref ref-type="bibr" rid="ref8 ref9">8, 9</xref>
        ].
      </p>
      <p>
        At the same time, despite the significant interest in the topic of MLOps from both scientists and
practitioners, there are still certain gaps and unresolved problems in this area. In particular, there are no
generally accepted standards and best practices for implementing MLOps, issues of integrating MLOps
with other approaches (DataOps, ModelOps, AIOps, etc.) are insuficiently researched, and there is a
need to develop new tools and platforms to automate MLOps processes [
        <xref ref-type="bibr" rid="ref10 ref11 ref5">5, 10, 11</xref>
        ].
      </p>
      <p>This work is aimed at solving the current problem of defining and analysing MLOps practices
necessary for the efective deployment of machine learning models. The basis for performing the work
is the need to systematize and generalize knowledge about MLOps practices, as well as the need to
develop recommendations for their implementation in organizations to increase the eficiency and
reliability of deploying machine learning models in production environments.</p>
      <p>According to the aim, the following main objectives of the study are defined:
1. Perform a meta-synthesis of systematic reviews to generalize knowledge about MLOps practices
necessary for the efective deployment of machine learning models.
2. Analyze the relationships between MLOps principles, processes, and practices.
3. Identify the most efective MLOps practices for deploying machine learning models.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Meta-synthesis of MLOps practices</title>
      <sec id="sec-2-1">
        <title>2.1. Main concepts of the study</title>
        <p>
          DevOps (Development &amp; Operations) is becoming increasingly widespread, and companies are
applying its methods in various fields [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ]. In this context, MLOps (Machine Learning &amp; Operations)
automates ML workflows, such as pipelines, applying DevOps practices (implementing continuous
integration/continuous deployment (CI/CD) for machine learning projects) [
          <xref ref-type="bibr" rid="ref12 ref6 ref8">12, 6, 8</xref>
          ].
        </p>
        <p>
          According to Calefato et al. [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ], key MLOps practices can be realized using GitHub Actions and CML
(Continuous Machine Learning). While some workflows automate ML tasks with GitHub Actions and
CML, production-grade, end-to-end MLOps pipelines appear rare in the analysed open-source GitHub
projects. Practices focus more on reporting and metrics rather than retraining or deployment.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Research methodology</title>
        <p>Systematic reviews can provide syntheses of the state of knowledge in a field, from which future
research priorities can be identified; they can address questions that otherwise could not be answered
by individual studies; they can identify problems in primary research that should be rectified in future
studies; and they can generate or evaluate theories about how or why phenomena occur [13, p. 1]. The
main aim of a systematic review is facilitating evidence-based decision-making [13, p. 6]. The main
diference between systematic and literature reviews is the “place of idea”. The literature review can be
idea-driven; thus, all sources can be selected to confirm some idea. Indeed, the systematic review is the
scientific method; instead of ideas, we operate with research questions and hypotheses. As a result, the
systematic review can produce new evidence-based knowledge.</p>
        <p>
          On 23 February 2024, we proceeded with a search request to the Scopus database by article title:
TITLE ( ( systematic OR review OR survey ) AND mlops )
5 documents were found (table 1), of which 3 [
          <xref ref-type="bibr" rid="ref5 ref6 ref8">6, 8, 5</xref>
          ] relate to systematic reviews.
        </p>
        <sec id="sec-2-2-1">
          <title>Continuation of table 1</title>
          <p>
            R. Cohen, Digital Strategy, Machine As part of a digital strategy, machine learning (ML) has become a
comLearning, and Industry Survey of MLOps, mon toolset and capability across many businesses. However, the
opin: Digital Strategies and Organizational erational aspects of machine learning (MLOps) are often overlooked
Transformation, 2023, pp. 137–150. for ML projects until they are already installed and being executed in
URL: https://tinyurl.com/33z6zpd3. the business environment. This chapter provides a review of MLOps
doi:10.1142/9789811271984_0008. [
            <xref ref-type="bibr" rid="ref15">15</xref>
            ] products and vendors to give data scientists the ability to set up the
appropriate ML infrastructure in a proactive manner.
          </p>
          <p>
            J. Diaz-de Arcaya, A. I. Torre-Bastida, G. Diaz-de Arcaya et al. [
            <xref ref-type="bibr" rid="ref5">5</xref>
            ] analyze the challenges, opportunities, and
Zárate, R. Miñón, A. Almeida, A Joint perspectives of implementing MLOps and AIOps. The authors
anaStudy of the Challenges, Opportunities, lyzed the open issues, opportunities, and trends faced by organizations
and Roadmap of MLOps and AIOps: A when implementing MLOps and AIOps, the frameworks and
architecSystematic Survey, ACM Comput. Surv. tures, as well as the fields of their use. The systematic review of 93
56 (2023) 84. doi:10.1145/3625289. studies provided an opportunity to identify: 1) successful
implementation of artificial intelligence projects requires a collaborative culture
and a combination of software engineering, data science, and DevOps
skills; 2) containerization, data and model versioning, FaaS
(Functionas-a-Service) and serverless architectures are useful for supporting the
MLOps/AIOps lifecycle; 3) monitoring the environment is important
for retraining and redeploying components; 4) AIOps is used
predominantly in complex environments, such as 5G and 6G technologies, while
          </p>
          <p>MLOps is more common in traditional industrial environments.</p>
          <p>
            The papers by Haertel et al. [
            <xref ref-type="bibr" rid="ref14">14</xref>
            ] and Cohen [
            <xref ref-type="bibr" rid="ref15">15</xref>
            ] are not systematic reviews, but the results obtained in
these works were taken into account when performing the meta-synthesis [
            <xref ref-type="bibr" rid="ref16">16</xref>
            ] for combining (appendix
B) and thematic analysis of the results obtained in the systematic reviews.
          </p>
          <p>
            The meta-synthesis was performed according to Chrastina [17, pp. 123-125]:
1. Defining the subject of research : MLOps practices for the efective deployment of models.
2. Identifying relevant sources: systematic literature reviews [
            <xref ref-type="bibr" rid="ref5 ref6 ref8">6, 8, 5</xref>
            ] and a review of MLOps products
and providers [
            <xref ref-type="bibr" rid="ref15">15</xref>
            ].
3. Thorough study to determine the common time period for the analyzed works, similarities and
diferences: in the aim, research questions, sources, inclusion, exclusion, and quality criteria,
definitions of MLOps and MLOps stages.
4. Defining the relationship between works through identifying and grouping key topics.
5. Mutual translation of results of diferent works through defining common terminology, explaining
contradictions in the results from diferent works, and generalizing results from diferent works.
6. Synthesis of results.
          </p>
          <p>7. Publication of the meta-synthesis.</p>
        </sec>
      </sec>
      <sec id="sec-2-3">
        <title>2.3. Thorough study and defining the relationship between works</title>
        <p>
          2.3.1. Distribution of reviews by year
Two works [
          <xref ref-type="bibr" rid="ref6 ref8">6, 8</xref>
          ] refer to 2022, one [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] – to 2023. At the same time, the sources analyzed in [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] are
limited to 2020, in [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] – to 2021, and in [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] – to 2023. In addition, the work [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] mentions the work [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] as
a previous one.
2.3.2. Review objectives
The aim of the meta-synthesis of the review objectives [
          <xref ref-type="bibr" rid="ref5 ref6 ref8">6, 8, 5</xref>
          ] was to identify common and diferent
aspects regarding the general focus and tasks of these studies.
        </p>
        <p>
          Common aspects of the objectives of the considered reviews are:
1. All reviews [
          <xref ref-type="bibr" rid="ref5 ref6 ref8">6, 8, 5</xref>
          ] are aimed at researching and generalizing knowledge about the MLOps
methodology, its practices, tools, and challenges.
2. The reviews [
          <xref ref-type="bibr" rid="ref6 ref8">6, 8</xref>
          ] aim to identify and analyze MLOps tools used to automate the development
and deployment processes of machine learning models.
3. The reviews [
          <xref ref-type="bibr" rid="ref5 ref8">5, 8</xref>
          ] seek to provide an understanding of the general state of MLOps practices
implementation in industry and academia.
        </p>
        <p>
          Distinct aspects of the review objectives are:
1. The review [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] focuses more on identifying and analyzing the functional capabilities of MLOps
tools for creating machine learning pipelines.
2. The review [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] pays attention to a wider range of MLOps aspects, such as practices, roles, maturity
models, challenges, in addition to tools.
3. The review [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ], in addition to the MLOps methodology, also considers the related concept of
AIOps. More attention is paid to highlighting the opportunities, challenges, and future trends in
both areas.
        </p>
        <p>
          Thus, despite some diferences in focus and breadth of coverage, all the considered reviews are
united by a common goal – to investigate and generalize knowledge about the MLOps methodology, its
practical application, tools, challenges, and state of implementation to promote further development of
this area.
2.3.3. Review research questions
The aim of the meta-synthesis of the research questions of the reviews [
          <xref ref-type="bibr" rid="ref5 ref6 ref8">6, 8, 5</xref>
          ] was to identify
common and diferent aspects regarding the main directions of research within the study of the MLOps
methodology.
        </p>
        <p>
          Common aspects of the research questions of the considered reviews are:
1. All reviews [
          <xref ref-type="bibr" rid="ref5 ref6 ref8">6, 8, 5</xref>
          ] contain questions about tools and platforms used to implement MLOps
practices, automate development processes, deployment, and monitoring of machine learning
models.
2. The reviews [
          <xref ref-type="bibr" rid="ref5 ref8">8, 5</xref>
          ] include questions regarding challenges and open problems faced by
organizations when implementing MLOps.
3. The reviews [
          <xref ref-type="bibr" rid="ref5 ref8">8, 5</xref>
          ] consider questions about opportunities and future trends in the field of MLOps.
Distinct aspects of the research questions of the reviews are:
1. The review [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] contains more specific questions about the functional capabilities and features of
        </p>
        <p>
          MLOps tools for creating machine learning pipelines.
2. The review [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] includes questions regarding the roles and responsibilities of specialists involved
in MLOps implementation, as well as maturity models for assessing the level of automation of
model deployment processes.
3. The review [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ], in addition to MLOps, also considers questions specific to the AIOps methodology
and pays attention to current and future areas of application of these approaches.
        </p>
        <p>
          Thus, the research questions of the considered reviews cover a wide range of MLOps aspects, from
tools and platforms to challenges, opportunities, and areas of application. Despite some diferences in
the focus of the questions, all reviews seek to explore the key components and factors that influence
the implementation and development of MLOps practices in organizations.
2.3.4. Review information sources
The aim of the meta-synthesis of the information sources of the reviews [
          <xref ref-type="bibr" rid="ref5 ref6 ref8">6, 8, 5</xref>
          ] was to identify common
and diferent aspects regarding the databases, search engines, and types of literature used to search for
relevant studies.
        </p>
        <p>
          Common aspects of information sources of the considered reviews are:
1. All reviews [
          <xref ref-type="bibr" rid="ref5 ref6 ref8">6, 8, 5</xref>
          ] used electronic databases of scientific publications to search for relevant
studies.
2. The reviews [
          <xref ref-type="bibr" rid="ref5 ref6">6, 5</xref>
          ] included both academic (peer-reviewed) and non-academic (“grey”) literature
sources such as blogs, websites, videos, code repositories, etc. in the search.
        </p>
        <p>
          Distinct aspects of information sources of the reviews are:
1. The review [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] used Google Scholar to search for scientific publications and regular Google search
for “grey” literature.
2. The review [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] limited the search to only academic databases such as ACM Digital Library, IEEE
        </p>
        <p>
          Xplore, ScienceDirect, and SpringerLink.
3. The review [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] used several databases (Scopus, arXiv, Springer, IEEE), but the main source was
the Scopus database from Elsevier.
        </p>
        <p>
          Thus, the considered reviews demonstrate diferent approaches to the selection of information
sources. Some studies [
          <xref ref-type="bibr" rid="ref5 ref6">6, 5</xref>
          ] include both academic and non-academic sources to obtain a more complete
picture of the practical application of MLOps. Others [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] focus exclusively on peer-reviewed scientific
publications. The choice of sources can influence the coverage and type of studies found, and hence the
results and conclusions of the reviews.
2.3.5. Criteria for including information sources in reviews
The aim of the meta-synthesis of the criteria for including information sources in the reviews [
          <xref ref-type="bibr" rid="ref5 ref6 ref8">6, 8, 5</xref>
          ]
was to identify common and diferent aspects regarding the requirements that studies must meet for
inclusion in the analysis.
        </p>
        <p>
          Common aspects of the inclusion criteria in the considered reviews are:
1. All reviews [
          <xref ref-type="bibr" rid="ref5 ref6 ref8">6, 8, 5</xref>
          ] included studies that directly relate to the topic of MLOps, its practices, tools,
and application.
2. The reviews [
          <xref ref-type="bibr" rid="ref6 ref8">6, 8</xref>
          ] considered studies that describe the experience, practices, architecture, or
implementation of MLOps tools and processes.
        </p>
        <p>
          Distinct aspects of the inclusion criteria in the reviews are:
1. The review [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] included studies that describe the components of the minimal MLOps lifecycle or
present the experience and opinions of experts regarding MLOps.
2. The review [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] additionally included studies that assess the maturity of MLOps processes, consider
roles and responsibilities in the ML model development lifecycle, and identify challenges in
developing and implementing MLOps solutions.
3. The review [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] had more general inclusion criteria, considering studies published from 2018 to
2023 that contain new ideas and are closely related to the topic of MLOps and AIOps.
        </p>
        <p>
          Thus, despite some diferences, the inclusion criteria in the considered reviews mainly focus on
studies that directly relate to MLOps, describe practical experience, tools, and processes, and consider
various aspects of the ML model development lifecycle. The reviews [
          <xref ref-type="bibr" rid="ref5 ref8">8, 5</xref>
          ] have somewhat broader
criteria, also including studies related to maturity assessment, roles, and MLOps challenges.
2.3.6. Criteria for excluding information sources from reviews
The aim of the meta-synthesis of the criteria for excluding information sources from the reviews [
          <xref ref-type="bibr" rid="ref5 ref6 ref8">6, 8, 5</xref>
          ]
was to identify common and diferent aspects regarding the characteristics of studies that lead to their
exclusion from the analysis.
        </p>
        <p>
          Common aspects of the exclusion criteria in the considered reviews are:
1. The reviews [
          <xref ref-type="bibr" rid="ref6 ref8">6, 8</xref>
          ] excluded studies that do not provide suficient details about the architecture,
implementation, or application of MLOps tools and processes.
2. The reviews [
          <xref ref-type="bibr" rid="ref5 ref8">8, 5</xref>
          ] excluded studies published in languages other than English.
        </p>
        <p>
          Distinct aspects of the exclusion criteria for information sources from the reviews are:
1. The review [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] excluded studies that promote commercial MLOps platforms without providing
details on their implementation or use.
2. The review [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] excluded studies that relate only to the application of ML models without
considering MLOps aspects, as well as short articles, posters, and studies without access to the full
text.
3. The review [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] excluded studies with an insuficient number of citations (depending on the year
of publication), as well as materials with limited access (by subscription) and articles published in
insuficiently reliable sources.
        </p>
        <p>
          Thus, the considered reviews apply diferent exclusion criteria to filter out studies that do not meet
their requirements. The common aspect is the exclusion of studies with insuficient descriptions of
MLOps processes and tools, as well as non-English publications. Diferences lie in additional criteria,
such as the exclusion of commercial platforms without technical details [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ], short articles and posters
[
          <xref ref-type="bibr" rid="ref8">8</xref>
          ], and materials with limited access and a low number of citations [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ].
2.3.7. Quality criteria for information sources in reviews
The aim of the meta-synthesis of the quality criteria for the reviews [
          <xref ref-type="bibr" rid="ref5 ref6 ref8">6, 8, 5</xref>
          ] was to identify common
and diferent aspects regarding the requirements for the quality and reliability of studies included in
the analysis.
        </p>
        <p>
          Common aspects of the quality criteria in the considered reviews are:
1. The reviews [
          <xref ref-type="bibr" rid="ref5 ref8">8, 5</xref>
          ] evaluated the quality of studies based on the completeness of the description
of the methodology, context, and results.
2. The reviews [
          <xref ref-type="bibr" rid="ref5 ref8">8, 5</xref>
          ] considered the presence of substantiated evidence and arguments to support
the study conclusions.
        </p>
        <p>
          Distinct aspects of the quality criteria in the reviews are:
1. The review [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] used quantitative indicators of popularity (number of stars on Github, views on
        </p>
        <p>
          YouTube) to assess the quality and relevance of “grey” literature.
2. The review [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] evaluated whether the study presents empirical results, not just expert opinions,
and whether the results are properly validated.
3. The review [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] used an extended set of criteria, including the presence of a comprehensive
literature review, verification of results on use cases, the number of research questions addressed,
the availability of open access, publication in high-impact journals, and the number of citations.
        </p>
        <p>
          Thus, the considered reviews apply diferent approaches to assessing the quality of studies. The
common aspect is the desire to include studies with a complete description of the methodology and
substantiated results. However, the specific quality criteria difer: from the use of popularity indicators
for “grey” literature [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ], to the assessment of the empirical nature of the results [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] and consideration of
bibliometric indicators such as journal impact factor and number of citations [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ].
        </p>
      </sec>
      <sec id="sec-2-4">
        <title>2.4. Mutual translation of results from diferent works and synthesis of results</title>
        <p>
          For the mutual translation of the results from diferent works, in addition to systematic reviews [
          <xref ref-type="bibr" rid="ref5 ref6 ref8">6, 8, 5</xref>
          ],
a review of MLOps products and providers [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] was involved.
2.4.1. Definition of MLOps
The aim of the meta-synthesis of MLOps definitions in the reviews [
          <xref ref-type="bibr" rid="ref15 ref5 ref6 ref8">6, 8, 5, 15</xref>
          ] was to identify common
and diferent aspects in the understanding and interpretation of this concept.
        </p>
        <p>
          Common aspects of MLOps definitions in the considered reviews are:
1. All reviews [
          <xref ref-type="bibr" rid="ref15 ref5 ref6 ref8">6, 8, 5, 15</xref>
          ] consider MLOps as a set of practices, principles, and processes for
automating and managing the lifecycle of machine learning models.
2. The reviews [
          <xref ref-type="bibr" rid="ref5 ref6">6, 5</xref>
          ] emphasize the use of approaches and practices from DevOps in MLOps, such
as continuous integration, delivery, and monitoring.
3. The reviews [
          <xref ref-type="bibr" rid="ref15 ref8">8, 15</xref>
          ] emphasize the role of MLOps in operationalizing machine learning solutions
and transferring them to industrial operation.
        </p>
        <p>
          Distinct aspects of MLOps definitions in the reviews are:
1. The review [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] focuses more on the technical aspects of MLOps, such as model lifecycle
management, pipeline automation, and performance monitoring.
2. The review [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] considers MLOps as a set of practices specifically for operationalizing data science
solutions.
3. The review [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] emphasizes the use of software engineering and machine learning principles in
        </p>
        <p>
          MLOps to create model-based products.
4. The review [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] considers MLOps as a separate area that focuses on automating the machine
learning model lifecycle as part of companies’ digital strategies.
        </p>
        <p>
          Thus, despite some diferences in emphasis and wording, all the considered reviews define MLOps
as an approach for managing, automating, and operationalizing the processes of developing,
deploying, and supporting machine learning models based on practices from software
engineering and DevOps. MLOps is a key component for the successful implementation of machine
learning solutions in an industrial environment.
2.4.2. Stages of the MLOps workflow
The aim of the meta-synthesis of the stages of the MLOps workflow in the reviews [
          <xref ref-type="bibr" rid="ref15 ref5 ref6 ref8">6, 8, 5, 15</xref>
          ] was
to identify the most common steps in the lifecycle of developing and implementing machine learning
models.
        </p>
        <p>
          Common stages of the MLOps workflow in the considered reviews are:
1. All reviews [
          <xref ref-type="bibr" rid="ref15 ref5 ref6 ref8">6, 8, 5, 15</xref>
          ] include the stages of data collection and processing, model development
and training, and model deployment in the working environment.
2. The reviews [
          <xref ref-type="bibr" rid="ref15 ref5 ref6">6, 5, 15</xref>
          ] highlight the stage of monitoring the performance and degradation of
deployed models as an important part of the MLOps workflow.
3. The reviews [
          <xref ref-type="bibr" rid="ref15 ref6">6, 15</xref>
          ] include the stage of retraining models based on new data or on a schedule as
part of the MLOps lifecycle.
        </p>
        <p>
          Distinct aspects of the MLOps workflow stages in the reviews are:
1. The review [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] provides a detailed breakdown of the workflow stages, including the steps of data
extraction, analysis, cleaning, and transformation, as well as model validation.
2. The review [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] focuses less on detailed stages and more on general MLOps functions, such as
data collection, transformation, model training, and implementation.
3. The review [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] groups stages into broader categories, such as data management, distributed
training, deployment, and monitoring.
4. The review [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] additionally highlights the stages of generating predictions and managing models
and data as part of the MLOps workflow.
        </p>
        <p>
          Thus, despite diferent levels of detail and grouping, the considered reviews demonstrate general
consistency regarding the main stages of the MLOps workflow. These stages cover the entire lifecycle
of machine learning models, from data collection and processing to deployment, monitoring, and
retraining of models. Diferences in the presentation of stages reflect diferent approaches to structuring
and describing the MLOps workflow.
2.4.3. Frameworks and architectures that facilitate MLOps implementation
The aim of the meta-synthesis of frameworks and architectures that facilitate MLOps implementation
in the reviews [
          <xref ref-type="bibr" rid="ref15 ref5 ref6 ref8">6, 8, 5, 15</xref>
          ] was to identify the most common and efective approaches and technologies
in this area.
        </p>
        <p>
          Common frameworks and architectures that facilitate MLOps implementation, according to the
considered reviews, are:
1. The reviews [
          <xref ref-type="bibr" rid="ref5 ref6 ref8">6, 8, 5</xref>
          ] highlight open-source platforms and frameworks such as MLflow, Kubeflow,
and TensorFlow Extended (TFX) as key components of the MLOps ecosystem.
2. The reviews [
          <xref ref-type="bibr" rid="ref15 ref5 ref6">6, 5, 15</xref>
          ] emphasize the importance of using cloud computing platforms and services
such as AWS, Google Cloud, and Azure to deploy and scale MLOps solutions.
3. The reviews [
          <xref ref-type="bibr" rid="ref5 ref8">8, 5</xref>
          ] note that architectures based on containerization (e.g., using Docker) and
container orchestration (e.g., using Kubernetes) are key to ensuring portability and scalability of
MLOps solutions.
        </p>
        <p>
          Distinct aspects of the considered MLOps frameworks and architectures in the reviews are:
1. The review [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] additionally highlights MLOps pipeline orchestration platforms such as Apache
        </p>
        <p>
          Airflow, Jenkins, and Polyaxon.
2. The review [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] additionally mentions tools such as Kubeflow, Polyaxon, Comet.ml, Kafka-ML,
        </p>
        <p>
          MLModelCI for managing pipelines and deploying models.
3. The review [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] considers a broader range of architectural approaches, including the use of edge
computing, serverless computing, and event-driven architectures.
4. The review [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] focuses primarily on proprietary platforms and solutions from commercial
providers such as Iguazio, Domino Data Lab, Comet, and Valohai.
        </p>
        <p>
          Thus, there are many frameworks and architectural approaches that facilitate MLOps implementation,
from open platforms and libraries to commercial solutions and cloud services. Key factors are support
for automation, scalability, portability, and integration with existing systems and tools. The choice of
appropriate frameworks and architectures depends on the specific requirements and constraints of the
organization, as well as the level of maturity of its MLOps processes.
2.4.4. MLOps tools for creating machine learning pipelines and operationalizing models
The aim of the meta-synthesis of MLOps tools for creating machine learning pipelines and
operationalizing models in the reviews [
          <xref ref-type="bibr" rid="ref15 ref5 ref6 ref8">6, 8, 5, 15</xref>
          ] was to identify the most popular and functional tools in this
area.
        </p>
        <p>
          Common MLOps tools mentioned in the considered reviews are:
1. The reviews [
          <xref ref-type="bibr" rid="ref6 ref8">6, 8</xref>
          ] highlight MLflow as a popular open-source platform for managing the lifecycle
of machine learning models, experiments, and deployment.
2. The reviews [
          <xref ref-type="bibr" rid="ref15 ref6">6, 15</xref>
          ] mention cloud platforms from major providers such as AWS SageMaker,
        </p>
        <p>
          Google Cloud AI Platform, Azure Machine Learning, as tools for operationalizing models.
3. The reviews [
          <xref ref-type="bibr" rid="ref5 ref8">8, 5</xref>
          ] note that containerization tools such as Docker and orchestration tools such
as Kubernetes are often used to deploy models.
        </p>
        <p>
          Distinct aspects of the considered MLOps tools in the reviews are:
1. The review [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] provides a detailed list of tools for diferent stages of the MLOps pipeline, including
orchestration platforms (Apache Airflow, Jenkins, Kubeflow, Polyaxon, Seldon Core, etc.) and
deployment (TensorFlow Extended).
2. The review [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] additionally mentions tools such as Kubeflow, Polyaxon, Comet.ml, Kafka-ML,
        </p>
        <p>
          MLModelCI for managing pipelines and deploying models.
3. The review [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] focuses more on general categories of tools, such as experiment management
systems, data and model versioning, and infrastructure automation.
4. The review [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] details the functionality of popular commercial MLOps platforms such as Iguazio,
        </p>
        <p>Domino Data Lab, Comet, Valohai, etc. (table 2)</p>
        <p>
          Thus, there is a wide range of MLOps tools for creating machine learning pipelines and
operationalizing models, from open platforms such as MLflow, to commercial solutions from cloud providers and
specialized companies. The choice of specific tools depends on the needs and scale of the organization,
as well as compatibility with the existing technology stack.
2.4.5. Main features ofered by MLOps tools
The aim of the meta-synthesis of the main features ofered by MLOps tools in the reviews [
          <xref ref-type="bibr" rid="ref15 ref5 ref6 ref8">6, 8, 5, 15</xref>
          ]
was to identify the key capabilities and components of these tools.
        </p>
        <p>
          Common features of MLOps tools, highlighted in the considered reviews, are:
1. The reviews [
          <xref ref-type="bibr" rid="ref5 ref6 ref8">6, 8, 5</xref>
          ] note that MLOps tools usually provide capabilities for tracking experiments,
versioning models and data.
2. The reviews [
          <xref ref-type="bibr" rid="ref15 ref6 ref8">6, 8, 15</xref>
          ] emphasize the importance of automation and orchestration features of
        </p>
        <p>
          MLOps workflows, such as model training and deployment pipelines.
3. The reviews [
          <xref ref-type="bibr" rid="ref15 ref5 ref6">6, 5, 15</xref>
          ] indicate the presence in MLOps tools of components for monitoring the
performance and degradation of deployed models.
        </p>
        <p>
          Distinct aspects of MLOps tool features, considered in the reviews, are:
1. The review [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] provides a detailed classification of features into three categories:
a) general features related to all stages of the MLOps pipeline:
• open source support;
• scalability and elasticity;
• extensibility;
• cloud-agnostic or cloud environment support;
• metadata management;
• continuous integration and delivery (CI/CD);
• user interfaces: graphical (GUI), command line (CLI), application programming interface
(API);
b) data management features:
• real-time data streaming;
• data storage;
• data analysis, cleaning, and transformation;
• data monitoring;
• metadata management;
• providing data access via API;
c) model management features:
• experiment tracking and model versioning;
• model registry;
• automatic hyperparameter optimization;
• model testing (A/B testing);
• anomaly and model drift detection;
• model performance monitoring;
• model metadata management;
• model deployment via API.
        </p>
        <p>
          • support for various machine learning libraries and frameworks;
2. The review [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] additionally highlights features such as automatic hyperparameter optimization
of models and mobility support for deployment in diferent environments.
3. The review [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] notes the importance of integrating MLOps tools with existing systems and
supporting collaborative work of teams.
4. The review [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] details the features of commercial MLOps platforms:
• model development: environment for data analysis, feature development, training, and model
experiments;
• operationalization of model training: creating reproducible pipelines for model training and
testing;
• continuous model training: automatic support for the frequency of model retraining based
on schedule, events, or ad-hoc requests;
• model deployment: packaging, testing, and deploying trained models in the production
environment;
• generating predictions: providing predictions or classifications in real-time or batch
processing mode;
• monitoring model performance: tracking the eficiency and degradation of models, warning
about the need for retraining;
• data and feature management: support for storing, processing, and accessing data and
generated features.
        </p>
        <p>
          Thus, MLOps tools provide a wide range of features to support the lifecycle of machine learning
models, with a focus on automation, experiment tracking, versioning, monitoring, and model deployment.
Some tools ofer more specialized features, such as hyperparameter optimization or data management.
The choice of a tool with an appropriate set of features depends on the specific needs and goals of the
organization in the field of MLOps.
2.4.6. Ways of deploying machine learning models in production environments
The aim of the meta-synthesis of ways of deploying machine learning models in production
environments in the reviews [
          <xref ref-type="bibr" rid="ref15 ref5 ref6 ref8">6, 8, 5, 15</xref>
          ] was to identify the most common and critical practices in this
area.
        </p>
        <p>
          Common ways of deploying machine learning models in production environments, according to the
considered reviews, are:
1. The reviews [
          <xref ref-type="bibr" rid="ref15 ref5 ref6">6, 5, 15</xref>
          ] note that models are often deployed using container technologies such as
        </p>
        <p>
          Docker, which provides model mobility and isolation.
2. The reviews [
          <xref ref-type="bibr" rid="ref15 ref6 ref8">6, 8, 15</xref>
          ] indicate the prevalence of deploying models in cloud environments using
platforms and services from major providers such as AWS, Google Cloud, and Azure.
3. The reviews [
          <xref ref-type="bibr" rid="ref5 ref6">6, 5</xref>
          ] note that models are often deployed as web services using REST API or other
protocols to provide access to predictions in real-time.
        </p>
        <p>
          Distinct aspects of the considered ways of deploying models in the reviews are:
1. The review [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] additionally describes deploying models using orchestration platforms (Apache
Airflow, Jenkins, Kubeflow, MLflow, Polyaxon, Seldon Core, Valohai) to provide automatic scaling
and container management.
2. The review [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] notes that some MLOps tools, such as MLflow, Kubeflow, and Kafka-ML, have
built-in capabilities to facilitate model deployment in diferent environments.
3. The review [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] considers deploying models not only in the cloud but also on edge devices using
specialized frameworks such as TensorFlow Lite and Core ML.
4. The review [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] provides a detailed description of the main stages and features of the machine
learning model deployment process using CI/CD pipelines and support for diferent environments
using commercial MLOps platforms:
a) creating a CI/CD pipeline for models: MLOps platforms such as SageMaker, Azure ML, and
Databricks allow creating CI/CD pipelines to automate the model deployment process,
which includes the stages of building, testing, and deploying models, as well as tracking
artifacts and version management;
b) supporting diferent deployment environments : MLOps platforms typically support multiple
deployment environments, such as development, testing, and production environments;
models can be deployed in diferent environments using appropriate configurations and
access policies;
c) model deployment process:
• trained models are packaged in a standardized format (e.g., Docker container) along
with necessary dependencies;
• the model goes through testing and validation stages to ensure its correctness and
compliance with requirements;
• after successfully passing the tests, the model is deployed in the target environment
(when deploying in the production environment, additional security and monitoring
measures may be applied);
d) automation and orchestration of deployment: MLOps platforms use automation tools such
as Jenkins or GitLab CI/CD to ensure continuous integration and deployment of models,
which can be configured to automatically trigger on certain events, such as updating the
model code or the appearance of new data;
e) monitoring and management of deployed models: MLOps platforms provide tools for
monitoring the performance and metrics of deployed models in real-time: in case of problems
or model degradation, the platform can automatically initiate the process of retraining or
rolling back to the previous version of the model.
        </p>
        <p>
          Thus, the most common ways to deploy machine learning models in production environments are
the use of container technologies, cloud platforms and services, and the deployment of models as web
services. The choice of a specific approach depends on the requirements for latency, scalability, and
availability of models, as well as the existing infrastructure and ecosystem of tools in the organization.
2.4.7. Maturity models for assessing the level of automation in deploying machine learning
models
Lima et al. [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] refer to several maturity models for assessing the level of improvement in the development
process of machine learning solutions:
1. The maturity model proposed by Amershi et al. [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ], mentioned simultaneously in [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] and [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ].
        </p>
        <p>
          This model, based on the Capability Maturity Model (CMM) and Six Sigma methodology, checks
whether the activity: (1) has defined goals, (2) is consistently implemented, (3) is documented, (4)
is automated, (5) is measured and tracked, and (6) is continuously improved.
2. According to Dhanorkar et al. [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ], organizations can be classified into three levels of maturity for
developing machine learning solutions: (1) data-oriented, (2) model-oriented, (3) pipeline-oriented.
3. Lwakatare et al. [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ] describe five stages of improvement in development practices: (1) manual
process led by data science, (2) standardized process of experimental-operational symmetry,
(3) automated ML workflow process, (4) integrated software development and ML workflow
processes, and (5) automated and fully integrated CD and ML workflow process.
4. Akkiraju et al. [21] proposed an adaptation of the CMM model with the definition of five levels
of maturity for each assessed capability: (1) initial, (2) repeatable, (3) defined, (4) managed, and
(5) optimizing.
        </p>
        <p>
          All systematic reviews [
          <xref ref-type="bibr" rid="ref5 ref6 ref8">6, 8, 5</xref>
          ] indicate that the level of automation of MLOps processes is one
of the key factors in assessing the maturity of an organization in this area. Despite the fact that the
considered reviews do not provide an exhaustive description of MLOps maturity models, they emphasize
the importance of assessing the level of automation of model development, testing, and deployment
processes as a key factor in the maturity of an organization in this area. The adaptation of existing
software development maturity models to the specifics of MLOps can be an efective approach to
assessing and improving machine learning processes in an organization.
2.4.8. Roles and responsibilities identified in the activities of operationalizing machine
learning models
The aim of the meta-synthesis of roles and responsibilities identified in the activities of operationalizing
machine learning models in the reviews [
          <xref ref-type="bibr" rid="ref15 ref5 ref6 ref8">6, 8, 5, 15</xref>
          ] was to identify the key participants in the MLOps
process and their functions.
        </p>
        <p>
          Common roles and responsibilities identified in the considered reviews:
1. All reviews [
          <xref ref-type="bibr" rid="ref15 ref5 ref6 ref8">6, 8, 5, 15</xref>
          ] mention the involvement of data scientists / data science researchers who
are responsible for developing, training, and experimenting with machine learning models.
2. The reviews [
          <xref ref-type="bibr" rid="ref5 ref6 ref8">6, 8, 5</xref>
          ] highlight the role of data engineers / data providers who are involved in
extracting, processing, transforming, and ensuring the quality of data for model training.
3. The reviews [
          <xref ref-type="bibr" rid="ref15 ref5 ref6">6, 5, 15</xref>
          ] note the importance of DevOps engineers, ML/MLOps engineers, and
software engineers in operationalizing models, automating deployment processes, creating pipelines,
and managing environments.
4. The reviews [
          <xref ref-type="bibr" rid="ref5 ref6">6, 5</xref>
          ] emphasize the role of managers, leadership, and business stakeholders in
defining model requirements for deployment, decision-making, and supporting the MLOps
strategy.
5. The reviews [
          <xref ref-type="bibr" rid="ref5 ref6">6, 5</xref>
          ] emphasize the role of managers, leadership, and business stakeholders in
defining model requirements, decision-making, and supporting the MLOps strategy.
        </p>
        <p>
          Distinct aspects of roles and responsibilities, considered in the reviews:
1. The review [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] additionally highlights the roles:
• domain specialist has deep knowledge of the subject area, plays an important role in obtaining
data and validating results);
• computational scientist/engineer has high technical skills to prepare the environment for the
operation of machine learning models;
• ML scientist/engineer is responsible for designing new machine learning models, has in-depth
knowledge of statistics and ML algorithms;
• provenance specialist manages the supply of data in the lifecycle of developing machine
learning solutions, has knowledge of both the subject area and machine learning;
• manager assesses models before their publication;
• application developer develops applications in which the created models will operate;
• deployment lead assesses aspects related to infrastructure components when deploying ML
models to production.
2. The review [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] mentions the role of subject matter experts in labeling data in specific domains.
        </p>
        <p>
          Thus, despite some diferences in the detail of roles, the considered reviews recognize the need to
involve specialists from diferent areas – software development, data engineering, machine learning,
subject matter experts, and management, for the successful operationalization of machine learning
models. Close collaboration and communication between these roles is critical for implementing MLOps
practices in organizations (figure 1).
2.4.9. Challenges encountered in deploying machine learning models in production
environments
The aim of the meta-synthesis of challenges encountered in deploying machine learning models in
production environments in the reviews [
          <xref ref-type="bibr" rid="ref15 ref5 ref6 ref8">6, 8, 5, 15</xref>
          ] was to identify the most common and critical
problems in this area. In [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] and [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ], specific challenges are not explicitly listed, but they can be
determined indirectly based on the discussion of MLOps and automation of machine learning pipelines
in [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] and the description of the various stages of MLOps and the need for appropriate tools in [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ].
        </p>
        <p>
          Common challenges identified in the considered reviews:
1. The reviews [
          <xref ref-type="bibr" rid="ref15 ref5 ref6 ref8">6, 8, 5, 15</xref>
          ] note the complexity of managing the machine learning model lifecycle,
including versioning, tracking, and reproducibility of models and data, as well as the problem of
ensuring scalability and performance of models in real-world conditions with large amounts of
data and requests.
362–414
        </p>
        <p>DS
Data Scientist
(ML model development)</p>
        <p>BE
(ML infrastructure management)</p>
        <p>ML</p>
        <p>ML Engineer /
MLOps Engineer
(cross-functional management
of ML environment and assets:</p>
        <p>ML infrastructure,</p>
        <p>ML models,
ML workflow pipelines,
data Ingestion,
monitoring)</p>
        <p>SE</p>
        <p>{…}
(applies design patterns and</p>
        <p>coding guidelines)</p>
        <p>DE
Data Engineer</p>
        <p>(data management,
data pipeline management)</p>
        <p>DO
(Software engineer with DevOps skills,</p>
        <p>ML workflow pipeline orchestration,</p>
        <p>
          CI/CD pipeline management,
monitoring)
FFiigguurer1e: I3nt.ersRectoiolneosf raolens dandtrhesepoinrsibiinlittieesr(ascceorcdtinigoton[s2, pc. o5]n). tributing to the
MLOps paradigm
2. The reviews [
          <xref ref-type="bibr" rid="ref15 ref5 ref8">8, 5, 15</xref>
          ] point to challenges associated with monitoring and maintaining models in
the production environment, including detecting data drift and model performance degradation.
5
3. The reviews [
          <xref ref-type="bibr" rid="ref6 ref8">6, 8</xref>
          ] consider challenges related to integrating software development with the
        </p>
        <p>
          Amarchcinhelietarenicngtpuipreleine.and Workflow
On 4t.hTehebraevsiieswso[f6,t5h]ecoindsiednerticfhiaelledngpesrirnelcatiepdlteose,nscuorimng pdaotanseencutrsit,yaannddprriovalecys,wwhene
deploying machine learning models.
derive a generalized MLOps end-to-end architecture to give ML
5. The reviews [
          <xref ref-type="bibr" rid="ref15 ref5">5, 15</xref>
          ] indicate that the quality, availability, preparation, labeling, and integration of
researdacthaferorms diafernendt sopurrcaecsitsiatisoignnieficarnst chparlloenpgeetrhatgrueqiudiraesnacloet.of Ititmeiasnddreesopuirccetse,adnd in
Figurheig4hl.ight the problem of interwpreetindgeapndicextptlahineinwgtoherrkesfullots wofsm,oid.eelo.,petrhateionsteoqenude-unsecres
        </p>
        <p>Additionally,
and business stakeholders.
in which the different tasks are executed in the different stages. The</p>
        <p>
          Distinct challenges, considered in the reviews:
artifact was designed to be technology-agnostic. Therefore, ML
1. The review [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] emphasizes the need to automate all stages of the MLOps pipeline and integration
reseawrcithheexrissting asonftdware pdervaelcoptmiteinotnsyesrtesms ancdapnrocescsehs.oose the best-fitting
tech2.nTohleoregviieews[8a]nndotefsrthaempreobwlemorokfsselefcotirngtahnedimrannaegeindgsin.frastructure for deploying models,
Asincluding the choice between clou4d,and on-premises environamnenetsn.d-to-end process,
depicted in Figure we illustrate
from MLOps project initiation to the model serving. It includes (A)
the MLOps project initiation st3e76ps; (B) the feature engineering
pipeline, including the data ingestion to the feature store; (C) the
dat
wa
fea
Af
an
pip
the
cle
sci
en
ad
rul
ba
en
mo
req
dat
bui
ini
the
en
mo
the
an
of
un
fea
(fo
clo
(1
cle
req
ma
tra
fee
3. The review [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] notes the problems: a) the gap between software engineering and machine
learning skills – data scientists often do not understand the requirements of certain production
environments, and software developers do not have suficient machine learning skills; b) efective
distribution, parallelization, and orchestration of data and ML tasks; c) the diversity of computing
infrastructure.
        </p>
        <p>
          The considered reviews show that deploying machine learning models in production environments
is associated with a number of challenges, such as managing the model lifecycle, ensuring scalability
and performance, monitoring and maintaining models in real-world conditions. Addressing these
challenges requires an integrated approach that includes automation of MLOps processes, selection
of appropriate infrastructure, ensuring data security and privacy, and efective communication with
business stakeholders.
2.4.10. Open issues, challenges, and peculiarities of MLOps
The aim of the meta-synthesis of open issues, challenges, and peculiarities of MLOps in the reviews
[
          <xref ref-type="bibr" rid="ref15 ref5 ref6 ref8">6, 8, 5, 15</xref>
          ] was to identify the most relevant and promising areas of research and development in this
ifeld. In [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] and [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ], open problems, challenges, and peculiarities of MLOps are not directly discussed,
but they can be identified based on the analysis of MLOps tools and their capabilities in [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] and the
description of the components and functions of MLOps platforms [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ].
        </p>
        <p>
          Common open issues and challenges of MLOps, identified in the considered reviews:
1. The reviews [
          <xref ref-type="bibr" rid="ref15 ref5 ref6 ref8">6, 8, 5, 15</xref>
          ] indicate the need to develop methods and tools to ensure the
interpretability, reproducibility, and responsible use of machine learning models in the context of
MLOps.
2. The reviews [
          <xref ref-type="bibr" rid="ref15 ref5 ref6">6, 5, 15</xref>
          ] emphasize the importance of developing approaches to data management
in MLOps, including ensuring data quality, privacy, and security.
3. The reviews [
          <xref ref-type="bibr" rid="ref6 ref8">6, 8</xref>
          ] note the need to develop and implement MLOps standards and best practices
to ensure consistency and compatibility between diferent tools and platforms.
4. The reviews [
          <xref ref-type="bibr" rid="ref5 ref8">8, 5</xref>
          ] emphasize the importance of the human factor in MLOps, including the need
to ensure efective communication and collaboration between diferent roles and teams and the
training of qualified personnel with cross-functional skills in programming, data processing, and
operational activities.
        </p>
        <p>
          Peculiarities of MLOps, identified in the considered reviews:
1. The review [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] notes that MLOps should take into account the specicfis of the machine learning
model development process, which difers from traditional software development.
2. The review [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] considers MLOps in the context of the overall digital strategy of the organization
and emphasizes the need to align MLOps practices with business goals and needs.
        </p>
        <p>
          Thus, the considered reviews identify a number of open issues and challenges in MLOps, such as the
need to develop standards and best practices, ensure interpretability and responsible use of models,
and efectively manage data. Peculiarities of MLOps, such as the diference from traditional software
development, the importance of the human factor, and the need to integrate knowledge from diferent
ifelds, require consideration when implementing MLOps practices in organizations.
2.4.11. Opportunities, future trends, and areas of application of MLOps
The aim of the meta-synthesis of opportunities, future trends, and areas of application of MLOps in
the reviews [
          <xref ref-type="bibr" rid="ref5 ref6 ref8">6, 8, 5</xref>
          ] was to identify promising directions of development and potential areas where
MLOps practices can bring significant benefits. In [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ], they are not directly discussed, but they can be
determined indirectly based on the presented MLOps tools and their capabilities, it is possible to outline
some potential directions and trends.
        </p>
        <p>
          Opportunities and future trends of MLOps, identified in the considered reviews:
1. The reviews [
          <xref ref-type="bibr" rid="ref5 ref6 ref8">6, 8, 5</xref>
          ] note the potential for developing standardized MLOps platforms and tools
that will simplify and accelerate the implementation of machine learning models in production.
2. The reviews [
          <xref ref-type="bibr" rid="ref5 ref8">8, 5</xref>
          ] note the prospects for integrating MLOps with other approaches, such as
DataOps, ModelOps, and DevSecOps, to provide comprehensive management of the machine
learning model lifecycle.
3. The review [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] points to: a) significant opportunities for further academic research and
development due to the fact that MLOps is still at an early stage; b) the expectation of increasing
demand for MLOps tools and platforms with the spread of artificial intelligence solutions; c) the
emergence of new roles and competencies related to MLOps as the industry develops.
4. The review [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] points to: a) opportunities to apply MLOps practices in the context of distributed
and federated model training, which will allow eficient use of decentralized data; b) involving
business units and training leadership in MLOps principles; c) using hardware platforms such as
FPGA and IoT to improve performance and privacy.
        </p>
        <p>
          Current and future areas of MLOps application, identified in the considered reviews:
1. The reviews [
          <xref ref-type="bibr" rid="ref5 ref6 ref8">6, 8, 5</xref>
          ] note that MLOps is already actively used in industries such as finance,
healthcare, commerce, marketing, and manufacturing, where machine learning models are used
to solve real business problems.
2. The review [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] points to the potential of applying MLOps in the field of IoT and edge computing,
where machine learning models can be deployed on resource-constrained devices, 5G and 6G
technologies, educational and scientific activities.
3. The review [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] notes the prospects for using MLOps in transportation and logistics.
        </p>
        <p>Thus, the considered reviews outline a number of opportunities and development trends for MLOps,
such as creating standardized platforms, applying in the context of distributed learning, and integrating
with other approaches to managing the lifecycle of data and models. Current and future areas of MLOps
application include a wide range of industries, from finance and healthcare to IoT and natural language
processing, which indicates a significant potential impact of this approach.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Analysis of MLOps practices</title>
      <sec id="sec-3-1">
        <title>3.1. Relationship between MLOps principles, processes, and practices</title>
        <p>MLOps is based on a set of principles [2, p. 3] and processes that ensure efective development,
deployment, and support of machine learning models (MLOps practices).</p>
        <p>MLOps principles define the fundamental foundations of designing machine learning pipelines:
• automation: maximum automation of all stages of the machine learning model lifecycle to reduce</p>
        <p>manual interventions and improve eficiency;
• reproducibility: ensuring the ability to reproduce the results of experiments and model deployment</p>
        <p>processes;
• collaboration: establishing efective collaboration and communication between diferent teams
involved in model development and implementation;
• continuous learning and improvement: regular updating of models based on new data and feedback,</p>
        <p>continuous improvement of MLOps processes;
• data governance: ensuring data quality, security, and confidentiality throughout the model
lifecy</p>
        <p>cle.</p>
        <p>MLOps processes define the sequence of actions for designing and implementing machine learning
pipelines:
1. Defining business goals and requirements : aligning the goals of developing machine learning</p>
        <p>models with the business strategy of the organization.
2. Data collection and preparation: collecting, cleaning, transforming, and enriching data for model</p>
        <p>training.
3. Model development and training: selecting algorithms, developing model architecture, training</p>
        <p>and validating models.
4. Model evaluation and testing: evaluating model performance on test data, conducting tests for</p>
        <p>reliability, security, and compliance with requirements.
5. Model deployment: packaging models with necessary dependencies, deploying in target
environ</p>
        <p>ments.
6. Model monitoring and maintenance: tracking model performance, identifying and resolving issues,</p>
        <p>updating models as needed.
7. Model lifecycle management: coordinating all stages of model development, deployment, and</p>
        <p>support, ensuring compliance with regulatory requirements.</p>
        <p>MLOps practices define the most efective methods and technologies for implementing machine
learning pipelines:
• basic MLOps practices include:
– continuous integration and delivery (CI/CD): automation of the processes of building, testing,</p>
        <p>and deploying machine learning models;
– model and data versioning: tracking changes in models and datasets, ensuring reproducibility</p>
        <p>of results;
– ML pipeline automation: creating automated workflows for data collection, processing,</p>
        <p>model training, and evaluation;
– model performance monitoring: tracking model quality metrics in the production
environ</p>
        <p>ment, detecting performance degradation;
– experiment management: organizing, tracking, and comparing diferent experiments with</p>
        <p>models and hyperparameters;
– model deployment: packaging models with necessary dependencies, deploying in diferent</p>
        <p>environments (cloud, edge, etc.);
– model lifecycle management: coordinating the processes of model development, testing,</p>
        <p>deployment, and monitoring to best ensure compliance with requirements;
• additional MLOps practices include:
– data security and privacy: ensuring the protection of data used for model training, compliance
with regulatory requirements;
– model explainability and interpretability: using methods and tools to understand and explain</p>
        <p>model behavior, especially in regulated industries;
– data quality management: monitoring and ensuring the quality of data used for model</p>
        <p>training and evaluation, detecting and handling anomalies;
– configuration management : versioning and managing configurations of environments where</p>
        <p>models are deployed, ensuring consistency across diferent environments;
– model deployment strategies: selecting and implementing appropriate deployment strategies;
– infrastructure automation: using Infrastructure as Code to automate provisioning and
man</p>
        <p>agement of infrastructure for model training and deployment;
– collaboration and communication: establishing efective collaboration between data science,</p>
        <p>development, operations, and business units;
– risk management and compliance: identifying and mitigating risks associated with the use</p>
        <p>of machine learning models, ensuring compliance with regulatory requirements.</p>
        <p>Figure 2 illustrates the relationships between key MLOps principles (light blue rectangles), main
processes (green rectangles), and common practices (orange rectangles). Arrows show how principles
influence processes, and processes, in turn, are implemented through specific practices. For example,
the principle of automation influences all MLOps processes, from goal definition to model lifecycle
management. The model development process is associated with practices such as versioning, pipeline
automation, experiment management, and model interpretability.
3.2. CI/CD
CI/CD (Continuous Integration/Continuous Delivery) is a key element/practice/implementation of
DevOps for automatic testing and deployment of code, data and models in a production environment
(figure 3) [ 7, p. 7]. In MLOps, it is extended to automate the process of developing and deploying ML
models, including the stages of building, testing, delivery, and deployment [2, pp. 3-4].</p>
        <p>The CI/CD process in MLOps includes the stages of build, test, delivery, and deploy [2, p. 4]. However,
unlike traditional CI/CD, MLOps may also have additional stages, such as model retraining.</p>
        <p>CI/CD in MLOps is part of the MLOps system architecture and provides fast feedback to developers
on the success or failure of certain stages, increasing overall productivity [2, pp. 3-4].</p>
        <p>Typical triggers for starting the CI/CD process in MLOps on GitHub are git push and pull_request
events [12, p. 4]. The events issue_comment, release, and schedule (on a schedule) can also be used.</p>
        <p>
          In the article Steidl et al. [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ], potential main triggers were also investigated, such as feedback systems
and alerts, scheduled orchestration service, traditional repository updates, and manual triggers. These
triggers start the execution of the pipeline, which consists of four stages: (1) data processing, (2) model
training, (3) software development, and (4) system commissioning. The data processing stage consists of a
repetitive end-to-end lifecycle of data-related tasks, such as preprocessing, quality assurance, versioning,
and documentation. The model training stage uses the results of data processing and illustrates model
development tasks such as model design, training, quality assurance, collecting metadata for model
improvement, version management, and documentation. After the pipeline completes model training,
the software development stage prepares the model for deployment through packaging, quality assurance
at the software level, and system versioning. In the final system commissioning stage, the model is
deployed in a specific environment using diferent deployment strategies and the system is monitored
[7, p. 21].
        </p>
        <p>A feature of the CI/CD process in MLOps is the need to version not only code, but also data and
models. This allows for reproducibility and the ability to roll back to previous versions [2, p. 4].</p>
        <p>
          Among the popular tools for implementing CI/CD in MLOps are Jenkins [2, p. 3] and GitHub Actions
[
          <xref ref-type="bibr" rid="ref12 ref2">2, 12</xref>
          ] and tools from cloud providers such as AWS CodePipeline, Azure DevOps Pipelines, etc.
        </p>
        <p>
          Figure 4 from the article by Kreuzberger et al. [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] presents an end-to-end MLOps architecture and
workflow with functional components and roles involved at each stage. Let’s consider in more detail
each of the zones and stages depicted in the figure:
1. A. MLOps Project Initiation. At this stage, the business stakeholder (BS) analyzes the problem
and defines the goal. The data scientist (DS) formulates the ML problem based on the business
goal. The necessary data is also determined and an initial analysis is performed by the data
engineer (DE) and data scientist (DS).
2. Data Engineering Zone. This zone includes two sub-stages:
• B1. Requirements for feature engineering pipeline. Rules for transformation, cleaning,
        </p>
        <p>and calculation of new features are defined.
• B2. Feature Engineering Pipeline. A data processing pipeline is implemented that
receives data from various sources, applies transformations, and loads it into a feature store
system.
3. C. Experimentation. At this stage, the data scientist (DS) performs data analysis, preparation,
and validation, as well as model training and validation. The best model is saved in the Model
Registry.
Literature rDevaineywlo O. Hanchuk et al. CEUR Workshop Proceedings
System Security for</p>
        <p>MLBSS</p>
        <p>Model extraction
attacks</p>
        <p>Model stealing</p>
        <p>attacks
Model-oriented
attacks</p>
        <p>Model adversarial</p>
        <p>attacks
Model trojan</p>
        <p>attacks
System-oriented
attacks
Platform attacks</p>
        <p>Model poisoning</p>
        <p>attacks
Model-reuse</p>
        <p>attacks
Model evasion</p>
        <p>attacks
System-level DNN</p>
        <p>attacks
Covert channel
attacks</p>
        <p>Data security and privacy is ensured by implementing various control and protection mechanisms at
holding local data samples [199]. The settings have a larger liability of exposing the attack surface to the adversary, in
each stage of MLOps. This includes, in particular, data encryption and anonymization, authentication
which the aadnvdearscacersys ccaonntmroal,ninipteuglarittey tchheectkraining,inacgtidvaittya minondiitfeorreinngt,sitnacgideesn.tInregspenonesrea,l,asthweeslluacscreesgsuolafrmseaccuhriintye learning
algorithms aausdsiutsmaensdtpheenienttreagtirointyteosftitnhge[t3r0a,ipnpin.4g-5d]a. ta and serving data. Availability of the training data is one main
concern for security. The goals of data-ormieani ntesdeactutraictyksccoamnpboenseunmtsmanardisperdacatsicteos: 1in.dMegLrOadpes tphreesmenatcehdinien l[e3a0r,ning model</p>
        <p>Figure 13 summarizes the
pp. 22-25].
performance; E2s.smenatinailpcuolmatpeonthenetpsroefdMicLtOiopns:outcome at test time.</p>
        <p>Review &amp; discussion:</p>
        <p>• problem definition includes understanding the problem and system requirements;</p>
        <p>Overall, the data-oriented attacks could be categorised into three major types of attacks, which are poisoning attacks,
backdoor attack•s daantda amdavneargseamriaelnattctoavcekrss. dWathailceoplloecistioonni,nlgabaetltiancgkasnhdavveerbificeaetniona;general term summarising most popular
attacks, in this •wmorokdeplociosonnstirnugctaiottnaacnkds rdeeprloeysmenetnsttihneclturdigesgemroledsesl speoleiscotinoinn,gbuaitltdaicnkgs,owphtiimleiztahteiobnaacnkddeovoarluaat-tacks is the
backdoor poisonintigona,ttaascwkse.llTarsigitgsedrleepslsoypmoiesnotniinngthaetttaacrkgestreenfevrirtoonsmuecncet;ssfully mislead the model and system behaviour
without manipu•lastyinstge mof mthaeindtaentaandcueringvomlvoedseml oinnfietorerinncgetshteagfuen.Fctoiornbiancgkodfotohreadtteapclokys,etdhseyastetamc.k will be activated when
the input dataSeccounrittayinprsacctriacfetseidn tMriLgOgpesr:, which could be pixel-level features in an image or designed characters in a
sequence. In Table. 2, we have collectively summarised 15 most recently exemplar studies to cover the modes targeting
1. At the problem definition stage:
on diferent development stages.</p>
        <p>• risk assessment;
For both triggerless poisoning attacks and backdoor poisoning attacks, the attacks mostly happen during the data
• threat modeling.
management stage when the attackers could easily obtain the access to the dataset. This assumption is made against
the centralised2m.Aacththineedlaetaarmnianngagmeomdeenlst s[t7a,g7e1: , 91, 161, 172, 182, 190], whilst one specific scenario is identified for the
distributed machine•ledaartnaiflonwg dmiaogdrealms.; In [129, 176], the federated learning-based systems are studied that the global
model is poisoned by• aSgTgRrIeDgaEtminegththoedoeldoggey nfoordtehrmeaotdcellausspificdaattieosnl;earned from malicious participants. The last attacking
stage for backdoor p•ociosnocneipntguaalttmaocdkesliinsgd;iscovered in [88], in which the model is retrained periodically at system
maintenance stage. G•ednaetraavllayl,idaatttiaocnk;s have impacts on the data integrity and availability.</p>
        <p>Although several•eadraltiaerlinwteorr;ks [120, 128, 143, 153] have identified the poisoning attacks in diferent applications,
one earliest work pr•esdeanttainvgeraificmatoiorne. realistic assumption by limiting the adversary’s capability and knowledge of the
system was led by Suciu et al. [172]. Neither the feature nor the algorithm knowledge was taken for granted as the
f the technical and data-related challenges. This, in turn, calls for adjusting the practices in the context of syste
omain and tecDhanniycloalOi.sHsaunecsh.uk et al. CEUR Workshop Proceedings</p>
        <sec id="sec-3-1-1">
          <title>3. At the model construction and deployment stage:</title>
          <p>
            Currently, the research and development of the state-of-the-practice for MLBSS are mostly focused on two topic
• identification of test adequacy criteria;
hich are technical deb•t taenstdinqguaaglaiitnystaasdsvuerrasnarciea.l iGnepnute;rally, quality assurance defines ‘ a planned and systematic patter
f all actions necessary •tofupzrzoivnigdteecahdneiqquueast.e confidence that the item or project conforms to established technical
requireents’ [
            <xref ref-type="bibr" rid="ref20">20</xref>
            ]. For technical debt, at its broadest, it was defined as ‘any side of the current system that is considered
4. At the system maintenance stage:
ub-optimal from a tech•nMicaLlmpoedrseplmecotniviteo’ri[n8g1,to1o7l7s](.MWLDhEilMe OthNe, dSeislfcCuhsescikoenr)o;f technical debt is related to the compromise
ecisions for the poor •sostfatwticaarnealdyesvise;lopment [93], the quality assurance dedicates to system quality ofline and in
roduction attached wi•thexspelvoerartaolrdyiastttiancckts;aspects, i.e., model quality, safety and fairness analysis [76].
          </p>
          <p>Technical debt, studie•deivnas[i1o7n7a]tftoacrkcsl;assical software systems, has recently been reviewed in MLBSS in order to identif
e specific patterns [ 1•4,d9a3ta, 1p6oi2s,o1n7in5g].aAttadcdkist;ional to typical technical debt at code level, the machine learning-base
ystems represent extra• tmraadnuea-oltfsesttoinbge; overcome for practical practices for the long term development, deployment
nd maintenance at th•e sdyynstaemmic laenvaelyl.sis.</p>
          <p>The data dependency debt was extensively discussed as a key contributor f
LBSS since it isThounse,soecfutrhitey ipnratectgicreasl icnoMmLpOopnseanretsinitnegMraLteBdSiSntoanadll misaignesnteagraeslloyf mthoerdeevdeilficoupmltetnot ldifeetceycctleas a result of
and include both general methods of security assessment and testing (threat modeling, static/dynamic
s unstable and underutilised characteristics [162]. Other technical debts, i.e., abstraction debt, configuration debt
analysis) and specialized approaches focused on the peculiarities of machine learning systems (testing
ata testing debagt,arinesptraoddvuercsiabriilailtyexdaembptl,eps,rdoecteecstsiomnaonf adgateampeonistondienbgt).and cultural debt, may also be accrued and deman</p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>3.10. Model explainability and interpretability</title>
        <p>
          ngoing collaborative eforts. Following in [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ], 21 primary studies were included to identify the actual technical deb
at have been investigated in recent works, in which four novel emerging debts of data, model, configuration and
        </p>
        <p>
          Explainability/interpretability of models is an important MLOps practice for ensuring transparency
thics were elaborated. Taking further actions towards measuring the technical debt and targeting to pay it of, a recent
and trust in machine learning models. Explainability refers to the ability to explain and understand the
anuscript submittdeedctiosioAnC-Mmaking processes of a machine learning model [10, p. 66]. This is especially important when
the model is used to make decisions that have significant consequences, such as in military operations
or law enforcement activities [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ].
        </p>
        <p>Explainability as the basis for trust in the ML project allows users to trust the prediction, which
increases transparency. The user can verify which factors contributed to certain predictions, introducing
an additional layer of accountability. The terms “explainability” and “interpretability” are often used
interchangeably, but for MLOPs, explainability is more than interpretability in terms of importance,
completeness, and fidelity of predictions or classifications (figure 14) [4, p. 63608].</p>
        <p>Individual case
explanation</p>
        <p>Model
accountability
Explainability</p>
        <p>Interpretability
Feature importance
explanation</p>
        <p>Human readability</p>
        <p>Explainable Artificial Intelligence (XAI) is a research direction that promotes explainable decision
making [4, p. 63609]. Explainability can be defined as the degree to which an observer can understand
the cause of a decision [32, p. 8]. An ML system is explainable when it is easier to identify causal
relationships between the inputs and outputs of the system. The more explainable a model is, the better
practitioners understand the internal business procedures that occur during model decision making. An
explainable model does not necessarily translate into a model that a human can understand (internal
logic or processes underlying it), but the explainability of the model allows the user to strengthen trust
in the predictions made by the deployed system [4, pp. 63614-63615].</p>
        <p>Various methods and tools can be used to achieve model explainability, such as:
• attribution-based methods (integrated gradients, saliency maps) or perturbation-based methods
(SHAP) [33, p. 3], which explain the model’s decision by assigning high scores to the most
influential input features;
• incorporating an attention mechanism into models, which allows focusing on the most relevant</p>
        <p>network states and inputs [33, p. 2];
• using a combined reward signal during training, which includes not only target metrics but also</p>
        <p>interpretability metrics [33, p. 4].</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.11. Data quality management</title>
        <p>
          Data quality management is an important practice in the MLOps workflow. According to Steidl et al.
[
          <xref ref-type="bibr" rid="ref7">7</xref>
          ], the data processing stage includes the full lifecycle of working with data, including pre-processing,
quality assurance, versioning and documentation [7, p. 21].
        </p>
        <p>In Haller’s book [9, pp. 77-84], data quality management is considered in the context of monitoring
and checking data in a production environment to ensure that the data used corresponds to reality.</p>
        <p>In the data-centric MLOps lifecycle (figure 15), data quality management is performed at the following
stages:</p>
        <p>1. Data collection – at this stage, data is created and obtained from various sources: data must be
relevant, complete, consistent and reliable.
2. Data quality assessment and cleaning – collected data undergoes thorough quality checks using
various metrics such as accuracy, completeness, consistency, timeliness, etc.; quality issues
are identified and resolved – incorrect values, missing values, duplicates, noise; data cleaning
techniques are applied to improve data quality.
3. Data augmentation and labeling – cleaned data is enriched with additional information and
labeled according to the target task; at this stage, the quality of data labeling is also controlled to
avoid errors and inaccuracies.
4. Data quality analysis – labeled data is analyzed for quality, its representativeness, class balance,
presence of outliers and anomalies are checked; adjustments are made as needed to improve data
quality.
5. Model training with quality control – based on quality data, ML model training takes place;
during experiments, model and data quality metrics are tracked to ensure stability and reliability
of results.
6. Model deployment with quality assurance – the model is deployed in the production environment
only after thorough testing on quality test data; measures are taken to maintain data quality in
the production environment.
7. Data and model quality monitoring – the deployed model and incoming data are constantly
checked for quality: quality metrics, presence of anomalies in data, distribution shift are tracked;
the model is retrained on new quality data as needed.</p>
        <p>To guarantee data quality throughout the entire lifecycle of their use in MLOps, some approaches to
data validation and verification need to be applied:
• data quality assessment by determining how suitable this data is for achieving business goals
(completeness, uniqueness, integrity, validity, accuracy, timeliness) [26, pp. 2-3];
• data validation in single and cross batches by comparing data characteristics with the expected
schema, as well as checking for data drift [7, p. 11];
• automated modular data tests based on a schema to identify errors in data and prevent them from
entering the model training stage [7, p. 10];
• versioning of data and related artifacts (processing procedures, metadata, etc.) for traceability,
reproducibility of results and compliance with regulatory requirements [7, p. 11];
• documenting data to an extent suficient to ensure model verifiability [7, p. 12].</p>
        <p>Timely detection and elimination of defects in data helps prevent wasting computational resources
on low-quality data [7, p. 11].</p>
        <p>In their work, Singh [26] present techniques for assessing the quality of “big data”, which help identify
datasets that may cause problems and unnecessary costs.</p>
        <p>To implement data quality management practices, frameworks such as TensorFlow Extended (TFX)
can be used, which have data validation components such as SchemaGen and ExampleValidator [7,
p. 19].</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.12. Configuration management</title>
        <p>Configuration files help create more robust software by moving all hardcoded variables into dedicated
locations that can be split up or organized at the developer’s discretion [34, p. 3].</p>
        <p>Godwin and Melvin [34] propose a template that supports two types of configuration files – a
config.py file that contains the template configuration and can include additional resources, such as
databases and spreadsheets, and JSON files for storing specific program variables, such as thresholds or
parameters. The config.py file is accessible through import statements, while JSON files are called from
disk at runtime [34, pp. 3-4].</p>
        <p>Yongqiang et al. [35] consider the use of a unified data model based on the YANG language for
unifying the description of configuration data and simplifying their management. Neptune Labs [36]
indicate that configuration management tools, such as Ansible (https://www.ansible.com/), Puppet
(https://www.puppet.com/) and Chef (https://www.chef.io/), can be used to automate configuration and
provisioning of MLOps platforms.</p>
      </sec>
      <sec id="sec-3-5">
        <title>3.13. Model deployment strategies</title>
        <p>Model deployment strategies:
• are implemented at the final stage of the MLOps process;
• require standardization, automation and encapsulation of models;
• are used for gradual transition of a new model to the production environment;
• rely on containerization.</p>
        <p>Peltonen and Dias [28] highlight the following benefits of using containers for model deployment
[28, p. 68]:
• model abstraction and process isolation by running multiple models in individual containers
representing their runtime dependencies;
• ability to create model-specific containers that meet their packaging requirements;
• assignment to diferent processors (CPU, Mobile GPU, etc.);
• ability to allocate a separate container to facilitate post-processing functions;
• a model repository and container storage can be used for fast deployment;
• provides a means of standardized deployment;
• almost all existing continuous development pipelines facilitate model deployment in the form of
containers;
• containers can be tailored to specific architectures of edge devices;
• Docker containers are a well-established industry standard;
• ease of rollback in case of failures.</p>
        <p>Gunny et al. [37] describe in detail the development cycle and model deployment strategy using
the DeepClean application as an example. A new version of the model is first deployed as a developer
version, undergoes validation in conditions similar to production, and only then replaces the current
model in the production environment. This uses a service architecture with NVIDIA Triton Inference
Server, which supports simultaneous placement of the developer version and the production version of
the model.</p>
        <p>The authors identify two main approaches to model deployment [37, pp. 11-12].</p>
        <p>WInotrhkesthraodiptioPnarlessceennatriaot,ieoacnh user manages their own resources and model versions.
Inconsistencies in libraries and dependencies, as well as model versions, lead to inconsistent results. Reduced
computational requirements for inference lead to underutilization of hardware resources, depicted
by green rectangles on each node (figure 16). More complex deployment scenarios require the use of
multiple networks, exacerbating existing issues.</p>
        <p>Git repository</p>
        <p>User node on
compute cluster
Local/shared
storage
Version of DL
software stack
Ops from different
DL frameworks
(b)
Figure 2: (a) A traditional deployment scenario in which individual users
Inconsistencies in libraries and39d8ependencies as well as model versions le
demands of inference lead to hardware under-utilization, represented
ser node on
ompute cluster
ocal/shared
torage
ersion of DL
oftware stack
ps from different
L frameworks
rio
Model as
point in
eam users
true test</p>
        <p>practices
yment via
table,
com</p>
        <p>Training Job
Training Job</p>
        <p>Cloud or
local model
repository
Containerized
inference service
gRPC inference
requests</p>
        <p>In the Inference-as-a-Service approach, a centralized service orchestrates models and provides unified
hich individual users manage their own software and hardware resources.</p>
        <p>interfaces for invoking models (figure 17). A centralized model store synchronizes all users and keeps
ell as modtheelmvueprtosidoatne.sPilpeelaindestsoendingRcPoCninsfiesretnecnerteqrueesstustlottsh.e Rseervdicueucseindg sctoanmdarpdiuzetdaAtPioIs nthaatl
abstract away the details of the inference execution itself. Inference is performed by a containerized
tilization, represented by green rectangles on each node. More complex
service that can eficiently schedule asynchronous model execution, maximizing hardware compute
networkscapuatbiillitiiezsiinnagpomrtaublletainpd lscealfabrlaemmanenwer.oInrtkhisbapapcrokacehn,cdonst,aienexriazactieonrballaowtisncrgeateinxgipsotrtianblge
dardizes ainndfiesorlaetendcmeodaelcerxoecsustioanlelnvuirsonemresntas.nd coordinates complex concurrent</p>
        <p>Choosing the right model deployment strategy allows minimizing operational costs, ensuring
consiscapacity in a way that is portable and scalable.</p>
        <p>tent model operation for all users, and facilitating model version monitoring and control.</p>
      </sec>
      <sec id="sec-3-6">
        <title>3.14. Infrastructure automation</title>
        <p>p. 3].
quired to</p>
        <p>users to pick and choose which libraries they need in order to
Infrastructure as code is an important MLOps practice that allows treating infrastructure as code for its
k like the reliable kanedeepficietnhtedierpldoyempelnotyanmdemnatnsagleimgehntt w[e3i8g, hp.t2.]I. nInftrhasitsruscetucrteiaounto,mwateionwreillaltebsrtioetflyhe
that look deploymdeentsacnrdibmeontihtoerinpgusrtapgoesinethoefMthLOepmsporosctesrse(figleuvrea1n8)t. Iltibenrsaurreiserse,liaabnledcrleeaativone omftohere
necessary environment for deploying machine learning models and automates infrastructure tasks [38,
detailed information to their documentation.</p>
        <p>Infrastructure as code involves describing infrastructure configuration in a declarative way in special
ifles (for e4x.a1m.1ple, inhYeArMmLe, sJS.OcNlfoorumdatbs)roerauskin.g Tspheceiaclllanoguudagbers(efoarkexlaimbprlae,rTyercroafnortma)inorstotoolso. ls
These filefsodrescprriboevthiseidoesniriendgstaKteuobfethrenineftraesstruccltuusrete[3r8s,pa.n4]d. virtual machines on
priDescribing infrastructure settings as code allows automating the process of its creation, modification
vate clouds and deploying workloads onto those computational
and management [38, p. 3]. This makes it possible to fully reproduce environments, quickly deploy
resourcerseasnoduavrociedsm. aWnuahl isleetuspuerproprosr[3t8,opn.4ly]. TchuisrrapepnrotlaycheisxaipsptrsopfroiarteGtoouosgelfoer Ccreloatuindg,
complex dynamic infrastructures, when high repeatability and consistency of environments is needed,
the intent of the library is to be written in such a way that the
user interface is agnostic to the actual cloud backend. Moreover, by
using Python contexts to deploy resources, we can ensure that any
399
resources are spun-down once jobs are complete so that
unnecesData collection
and preparation</p>
        <p>Data analysis
and exploration</p>
        <p>Model building
and training</p>
        <p>Model testing
and validation</p>
        <p>Model deployment
(Infrastructure automation)</p>
        <p>Infrastructure as Code</p>
        <p>Model retraining and updating</p>
        <p>Model
performance
monitoring
to increase reliability and reduce time costs for manual configuration [38, pp. 2, 4].</p>
        <p>The automated approach allows testing infrastructure as code, applying software development
practices to it (code review, versioning, etc.). This contributes to improving stability and security, allows
quickly tracking and resolving infrastructure issues [38, p. 4].</p>
        <p>To implement Infrastructure as code in Azure DevOps, tools such as Azure Resource Manager (ARM)
Templates, Terraform, Ansible, Chef are used [38, pp. 3-4]. They allow describing infrastructure as code
and automatically creating or modifying it.</p>
      </sec>
      <sec id="sec-3-7">
        <title>3.15. Collaboration and communication</title>
        <p>Collaboration and communication between various stakeholders is a key MLOps practice for the
successful implementation of machine learning and artificial intelligence projects in organizations.
MLOps emphasizes how cross-functional teams, such as data analysts, system operators, as well as data
and software engineers, collaborate through a harmonized process [7, p. 7].</p>
        <p>The essence of the collaboration and communication practice lies in establishing efective interaction
and information exchange between the various teams involved in the process of developing and
implementing ML models – the data science team, development team, operations team, and business
units. This practice is an important component of the development and implementation stage of the
MLOps workflow.</p>
        <p>
          As Kreuzberger et al. [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] point out, MLOps involves close collaboration between data science (machine
learning) teams, which are engaged in data preparation and model development, software development
engineers, who are responsible for integrating models into the production environment, and operations
teams, which ensure the deployment and support of models [2, p. 2]. Efective communication is
necessary at all stages of the ML model lifecycle, from defining business goals to monitoring models in
the production environment. Without well-established collaboration, the development of ML solutions
can be delayed, conflicts of interest and misunderstandings may arise between participants [39, p. 3].
        </p>
        <p>The diagram (figure 19) shows the main participants in the MLOps process and the directions of their
interaction:
1. The Data Science team prepares data and develops machine learning models.
2. The Development team integrates the developed models into software products.
3. The Operations team deploys models in the production environment and provides their support.
4. The Business team defines goals and requirements for ML solutions, and also receives results
from the Data Science and Operations teams.</p>
        <p>Data
preparation,</p>
        <p>model
development</p>
        <p>Defining goals and requirements</p>
        <p>Business
Collaboration
tools: Slack,
Trello, GitLab wiki</p>
        <p>Model
deployment
and support
Data Science</p>
        <p>Development</p>
        <p>Operations</p>
        <p>Integration of models into the product</p>
        <p>To ensure efective communication in MLOps practice, the following approaches and tools are used
[2, p. 4]:
• use of collaboration and knowledge sharing tools, such as Slack, Trello, GitLab wiki;
• regular meetings between teams to discuss status, problems and plans;
• clear definition of roles and areas of responsibility of process participants;
• use of version control systems (Git) for collaboration on code and models;
• automation of CI/CD processes to ensure transparency and reproducibility of development.</p>
      </sec>
      <sec id="sec-3-8">
        <title>3.16. Risk management and compliance</title>
        <p>Since ML models often make important decisions that afect people, risk management and compliance
is a critically important MLOps practice, the essence of which is to ensure compliance of developed ML
models and systems with regulatory requirements and standards (compliance) and manage potential
risks associated with their development and operation.</p>
        <p>
          This practice has a cross-cutting nature and manifests itself at diferent stages of the ML model
lifecycle. Risk management in the context of MLOps involves identifying and mitigating potential risks
associated with ML models, such as data bias, privacy breaches, model accuracy deterioration over
time, etc. Compliance means ensuring that models comply with regulatory requirements, for example,
regarding data protection [7, pp. 18-19]. The main condition for using this practice is the presence of
regulatory requirements or industry standards that the ML system must comply with. Examples can be
GDPR requirements for personal data protection or certifications in the healthcare industry [ 7, p. 5].
Steidl et al. [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] indicate that compliance aspects should be considered during data preparation, model
training and validation, as well as deployment and monitoring [7, p. 18].
        </p>
        <p>Methods of using this practice include regular data quality checks, testing models for bias and
discrimination, implementing access control and data encryption, documenting model architecture and
development process, monitoring model performance after deployment [7, pp. 11-12]. Yes, compliance
is achieved by [7, p. 18]:
• quality control and origin of data used for model training to avoid violation of regulatory
requirements;
• documenting and versioning models and data to ensure reproducibility of results and auditing;
• verifying models integrated into the production environment for compliance with requirements;
• continuous monitoring of deployed models for timely detection of potential violations or incorrect
behavior.</p>
        <p>
          The leading tools for ensuring risk management and compliance practice are version control systems
for tracking changes in data and model code, testing tools for identifying problems in models, monitoring
systems for tracking model accuracy in real time [7, pp. 12, 14]. At the same time, interviews with
practitioners conducted by Steidl et al. [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] revealed that ensuring compliance of ML systems in various
domains (for example, in healthcare) is a serious challenge due to the lack of established methodologies
and software tools. At the same time, achieving compliance is a mandatory condition for obtaining the
necessary certifications and permits from regulators.
        </p>
        <p>The diagram in figure 20 shows three main MLOps stages: working with training data, model training,
and system deployment (operationalization). At each stage there are blocks that indicate potential risks
and compliance measures. This diagram illustrates the cross-cutting nature of risk management and
compliance practice in MLOps, showing its presence and interconnections at each stage of the machine
learning model lifecycle.</p>
        <p>Figure 21 shows the relationships between the key principles, deployment process, and main MLOps
practices that are applied at the stage of machine learning model deployment. The principles of
automation and reproducibility influence the model deployment process, which is associated with
the following MLOps practices: CI/CD, model deployment, data security and privacy, configuration
management, model deployment strategies, and infrastructure automation.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Conclusions</title>
      <p>
        A meta-synthesis of systematic reviews [
        <xref ref-type="bibr" rid="ref5 ref6 ref8">6, 8, 5</xref>
        ] and a review of products and providers [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] was
performed in order to generalize knowledge about the implementation of MLOps practices for the
efective deployment of machine learning models. The main conclusions obtained as a result of the
meta-synthesis are as follows:
1. MLOps is an approach for managing, automating, and operationalizing the processes of developing,
deploying, and supporting machine learning models based on practices from software engineering
and DevOps. MLOps is based on a set of principles, processes, and practices that ensure efective
development, deployment, and support of machine learning models.
2. The main stages of the MLOps lifecycle include the following processes: data collection and
processing, model development and training, deployment, monitoring, and retraining of models.
3. Various frameworks and architectures are used to implement MLOps, such as open source
platforms (MLflow, Kubeflow, TensorFlow Extended), cloud computing platforms (AWS, Google
Cloud, Azure), containerization (Docker), and container orchestration (Kubernetes).
data
risks
model
risks
deployment
risks
data collection
model training
operationalization
data
compliance
      </p>
      <p>model
compliance
deployment
compliance
4. MLOps tools provide a wide range of features to support the machine learning model lifecycle,
with a focus on automation, experiment tracking, versioning, monitoring, and model deployment.
5. The most common ways to deploy machine learning models in production environments are the
use of container technologies, cloud platforms and services, and the deployment of models as
web services.</p>
      <p>Model deployment
Data security and privacy
Configuration management
Model deployment strategies</p>
      <p>Infrastructure automation
6. Adapted software development maturity models, such as CMM, can be used to assess the maturity
level of MLOps processes in organizations.
7. Successful MLOps implementation requires the involvement of specialists from diferent
areas - software development, data engineering, machine learning, subject matter experts, and
management.
8. The main challenges when deploying machine learning models in production environments are
managing the model lifecycle, ensuring scalability and performance, monitoring and maintaining
models in real-world conditions.
9. Open issues and challenges in MLOps are the need to develop standards and best practices, ensure
interpretability and responsible use of models, efectively manage data, integrate knowledge from
diferent fields.
10. The main opportunities and development trends of MLOps are the creation of standardized
platforms, application in the context of distributed learning, and integration with other approaches
to managing the lifecycle of data and models. Current and future areas of MLOps application
include a wide range of industries, from finance and healthcare to IoT and natural language
processing.</p>
      <p>The conducted meta-synthesis showed that MLOps is a promising approach for the efective
deployment of machine learning models in production environments, which requires further research and
development to address existing challenges and realize potential opportunities.</p>
      <p>Next, the key MLOps practices necessary for efective deployment of machine learning models were
analyzed. The main conclusions obtained as a result of the analysis are as follows:
1. MLOps is based on a set of principles, processes and practices that ensure efective development,
deployment and maintenance of machine learning models. The key principles of MLOps are
automation, reproducibility, collaboration, continuous learning and data governance.
2. The main MLOps practices include: continuous integration and delivery (CI/CD), model and data
versioning, ML pipeline automation, model performance monitoring, experiment management,
model deployment and lifecycle management.
3. Additional MLOps practices, such as data security and privacy, model explainability and
interpretability, data quality management, configuration management, model deployment strategies,
infrastructure automation, collaboration and communication, risk management and compliance,
are important for ensuring reliability, compliance with requirements and eficiency of MLOps
processes.
4. The application of MLOps practices allows automating and standardizing the processes of
development, deployment and maintenance of machine learning models, which increases the eficiency
and reliability of ML solutions in the production environment.
5. Successful implementation of MLOps practices requires the use of appropriate tools and platforms,
such as experiment management systems, data and model versioning, infrastructure automation
and monitoring tools, as well as establishing efective collaboration between the diferent roles
and teams involved in the process of developing and implementing machine learning models.</p>
      <p>The analysis showed that the application of MLOps practices is critically important for the successful
deployment of machine learning models in production environments. The implementation of these
practices allows increasing the eficiency, reliability and reproducibility of the processes of development
and operation of ML solutions, which is a necessary condition for their successful use in real business
problems.</p>
      <p>As a result of the comprehensive study, the MLOps practices necessary for the efective deployment
of machine learning models were identified and analyzed. The main conclusions obtained during the
study are as follows:
1. A meta-synthesis of systematic reviews was performed to generalize knowledge about MLOps
practices. The conducted meta-synthesis showed that MLOps is a promising approach for the
efective deployment of machine learning models in production environments, which requires
further research and development to solve existing challenges and realize potential opportunities.
2. A diagram of relationships between MLOps principles, processes and practices is proposed.</p>
      <p>This diagram illustrates the interconnections between the key principles, stages of the machine
learning model development and implementation process, as well as the main MLOps practices
that are applied at each stage.
3. The most efective MLOps practices for model deployment have been identified, which include:
continuous integration and delivery (CI/CD), model and data versioning, ML pipeline
automation, model performance monitoring, experiment management, model deployment and lifecycle
management, data security and privacy, model explainability and interpretability, data quality
management, configuration management, model deployment strategies, infrastructure
automation, collaboration and communication, risk management and compliance.</p>
      <p>The obtained results have both theoretical and practical significance. The theoretical significance
lies in the generalization and systematization of knowledge about MLOps practices necessary for the
efective deployment of machine learning models. The practical significance of the obtained results
lies in the possibility of their use by organizations for the implementation or improvement of MLOps
processes in order to increase the eficiency and reliability of machine learning model deployment in
production environments.</p>
      <p>Further research may be aimed at developing detailed recommendations for the implementation of
individual MLOps practices in organizations, creating new tools and platforms for automating and
managing the lifecycle of machine learning models, as well as studying the efectiveness of applying
MLOps practices in various industries and areas of machine learning model application.
Declaration on Generative AI: During the preparation of this work, the authors used Claude 3 Opus in order to: Text
Translation, Abstract drafting. After using this service, the authors reviewed and edited the content as needed and takes full
responsibility for the publication’s content.
Software Ecosystem for Deploying Deep Learning in Gravitational Wave Physics, in: Proceedings
of the 12th Workshop on AI and Scientific Computing at Scale Using Flexible Computing
Infrastructures, FlexScience ’22, Association for Computing Machinery, New York, NY, USA, 2022, p.
9–17. doi:10.1145/3526058.3535454.
[38] C. Vuppalapati, A. Ilapakurti, K. Chillara, S. Kedari, V. Mamidi, Automating Tiny ML Intelligent
Sensors DevOPS Using Microsoft Azure, in: 2020 IEEE International Conference on Big Data (Big
Data), 2020, pp. 2375–2384. doi:10.1109/BigData50022.2020.9377755.
[39] R. Sothilingam, V. Pant, E. S. K. Yu, Using i* to Analyze Collaboration Challenges in MLOps Project
Teams, in: A. Maté, T. Li, E. J. T. Gonçalves (Eds.), Proceedings of the 15th International iStar
Workshop (iStar 2022) co-located with 41th International Conference on Conceptual Modeling
(ER 2022), Virtual Event, Hyderabad, India, October 17, 2022, volume 3231 of CEUR Workshop
Proceedings, CEUR-WS.org, 2022, pp. 1–6. URL: https://ceur-ws.org/Vol-3231/iStar22_paper_1.pdf.</p>
    </sec>
    <sec id="sec-5">
      <title>A. Using the large language model Claude 3 Sonnet for analyzing systematic reviews</title>
      <p>The queries were created on 10.05.2024. A PDF file with the article text was added to each query.
A report according to the submitted plan was expected from the chatbot. The queries consisted of
universal and variable parts.</p>
      <p>Universal part of the query:
Using the added review article file as a data source,
write a detailed report on it according to the following plan.</p>
      <p>Variable parts of the query:
1. 1. Year of publication.</p>
      <p>2. Research objective (paper aim).</p>
      <p>3. Research questions.
2. 4. Information sources (databases, etc.)
5. Inclusion criteria.
6. Exclusion criteria.</p>
      <p>7. Quality criteria.
3. 8. MLOps definition (if any).</p>
      <p>9. MLOps workflow stages (if any).
4. 10. What frameworks and architectures facilitate MLOps?
5. 11. What MLOps tools can be used to build ML pipelines for</p>
      <p>Continuous Deployment? What tools are used in the
activities for operationalizing machine learning models?
6. 12. What are the main features offered by MLOps tools?
7. 13. How are machine learning models deployed in production</p>
      <p>environments?
8. 14. What maturity models are used to assess the level of</p>
      <p>automation in deploying machine learning models?
9. 15. What roles and responsibilities are identified in the
activities of operationalization of machine learning models?
10. 16. What challenges are encountered with regard to
deploying machine learning models in production
11. 17. What are the open issues, challenges and
particularities in MLOps?
18. What are the opportunities and future trends in MLOps?
What are the current and future fields in which MLOps is
thriving?</p>
    </sec>
    <sec id="sec-6">
      <title>B. Results of systematic reviews analysis</title>
      <p>2022</p>
      <p>
        Review by
Lima et al. [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]
      </p>
      <p>Review by</p>
      <p>
        Diaz-de Arcaya et al. [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]
2023
Table B.1
Results of systematic review analysis [
        <xref ref-type="bibr" rid="ref5 ref6 ref8">6, 8, 5</xref>
        ].
      </p>
      <p>Comparison Review by</p>
      <p>
        object Recupito et al. [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]
Year of publi- 2022
cation
Research
objective
      </p>
      <p>To identify tools that allow cre- Review of existing literature The main objective of this
sysating MLOps pipelines for con- to identify practices, standards, tematic literature review is to
tinuous deployment. roles, maturity models, chal- provide an understanding of the
To analyze the main character- lenges and tools used to auto- implementation of MLOps and
istics and functions of these mate activities of operational-AIOps methodologies in both
inMLOps tools to provide a com- izing machine learning mod- dustry and academia. The
auprehensive overview of their els into industrial operation thors seek to highlight the
chalvalue. (MLOps). lenges, opportunities and future
trends in these areas.</p>
      <p>Research ques- We answer the following main RQ1: How are machine learning Q1: What are the open issues,
tions research question: What tools models deployed in production challenges and particularities in
and capabilities enable develop- environments? MLOps and AIOps?
ers to create ML-enabled soft- RQ2: What maturity models are Q2: What are the opportunities
ware systems? used to assess the level of au- and future trends in MLOps?
Which was detailed into two tomation in deploying machine Q3: What are the opportunities
sub-questions: learning models? and future trends in AIOps?
RQ1. What MLOps tools can be RQ3: What roles and responsi- Q4: What platforms and
archiused to create machine learning bilities are defined in the activi- tectures facilitate MLOps and
pipelines for continuous deploy- ties of operationalizing machine AIOps?
ment? learning models? Q5: What are the current and
fuRQ2. What are the main fea- RQ4: What tools are used in ture fields in which MLOps and
tures ofered by MLOps tools? the activities of operationaliz-AIOps are thriving?
ing machine learning models?
RQ5: What challenges are
encountered when deploying
machine learning models in
production environments?</p>
      <p>Continued on next page</p>
      <sec id="sec-6-1">
        <title>Continuation of table B.1</title>
        <p>
          Comparison Review by Review by Review by
object Recupito et al. [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] Lima et al. [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] Diaz-de Arcaya et al. [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ]
Information Google Scholar – for searching The automated search was To search for relevant
artisources scientific literature, such as jour- conducted in the following cles, the authors used several
nals, books and dissertations. electronic research databases: databases and repositories,
inGoogle Search – for searching ACM Digital Library, IEEE cluding arXiv, Springer, IEEE.
so-called “gray” literature, such Xplore, ScienceDirect and However, the main source was
as blog posts, developer sites, SpringerLink. the Scopus database from
Elsewebinars, GitHub repositories vier, as it contains metadata and
and YouTube videos. abstracts of many publications.
Using both academic (white
literature) and professional (gray
literature) sources allowed the
authors to explore MLOps from
diferent perspectives –
theoretical and practical.
        </p>
        <p>Inclusion crite- The study discusses compo- Studies related to machine Articles published between
ria nents of a minimal end-to-end learning operations (MLOps) in 2018 and 2023, identified by
MLOps workflow. general. search queries, containing new
The study discusses MLOps Studies evaluating the lifecycle ideas and closely related to
practice or machine learning- of machine learning solutions. the topic of MLOps and AIOps
based applications. Studies related to maturity were included in the analysis.
The study relates to the imple- models of the machine learning
mentation of MLOps tool(s). process.</p>
        <p>The study describes experi- Studies analyzing roles and
ences, opinions, or practices re- responsibilities involved in the
garding MLOps pipelines. development and deployment
of machine learning solutions.</p>
        <p>Studies covering tools for
deploying machine learning
solutions.</p>
        <p>Studies identifying challenges
for the development and
deployment of machine learning
models.</p>
        <p>Exclusion cri- The study does not provide de- Studies not published in En- Publications not in English,
reteria tails on the design or implemen- glish. tracted publications,
publicatation of MLOps tool(s). Studies related to the applica- tions from irrelevant publishers,
The study only proposes the de- tion of machine learning mod- subscription materials without
sign of a certain component of els. access, and articles with
insufmachine learning pipelines. Short papers or posters. ifcient citations (depending on
The study does not provide or Studies not related to machine the year of publication) were
exreference details on machine learning operations. cluded.
learning automation. Studies whose content is not
acThe study refers to commercial cessible.
platforms that ofer MLOps ap- Articles not relevant to the
replications to promote their de- search questions.
velopment and deployment
services.</p>
        <p>Continued on next page</p>
      </sec>
      <sec id="sec-6-2">
        <title>Continuation of table B.1</title>
        <p>
          Comparison Review by Review by Review by
object Recupito et al. [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] Lima et al. [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] Diaz-de Arcaya et al. [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ]
Quality crite- The repository must have at Does the study report unam- Comprehensive literature
ria least 100 stars biguous findings based on ev- review and gap identification
The YouTube video must have idence and arguments? Verification of results on a use
been viewed at least 1000 times Does the study represent a re- case
search project and not an expert Number of research questions
opinion? addressed
Is the context being analyzed Publication under an open
fully described in the study? license
Are the research objectives Type of publication
(jourclearly defined? nal/other)
Are the research results prop- Publication in a high-impact
erly validated? journal
        </p>
        <p>Number of citations
MLOps defini- MLOps is a practice that helps A set of practices and princi- MLOps uses machine learning,
tion (if avail- model, develop and operate ples for operationalizing data DevOps and data engineering to
able) the machine learning lifecycle, science solutions, used to au- bring machine learning systems
drawing on DevOps principles tomate the deployment of ma- into production, facilitating the
and practices. chine learning models into an creation of machine products.</p>
        <p>operational environment
MLOps work- Data extraction for integration Data collection Data management
lfow stages (if Data analysis Data transformation Distributed training
available) Data cleaning, transformation Continuous training Deployment
and feature engineering to split Continuous model deployment Monitoring
data Presentation of results Retraining
Model training Monitoring of machine learning The need to manage the
lifecyModel validation to assess the solutions cle and key components of AI
quality applications using specialized
Model deployment in target en- platforms and tools is
emphavironments sized.</p>
        <p>Model monitoring
Frameworks Continuous training pipelines MLflow
and archi- deployed via CI/CD Kubeflow
tectures that Orchestration platforms Polyaxon
facilitate TensorFlow Extended (TFX) Comet.ml
MLOps imple- Machine learning cloud plat- Kafka-ML
mentation forms MLModelCI</p>
      </sec>
      <sec id="sec-6-3">
        <title>Continuation of table B.1</title>
        <p>
          Comparison Review by Review by Review by
object Recupito et al. [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] Lima et al. [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] Diaz-de Arcaya et al. [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ]
MLOps tools Machine learning cloud MLflow - this open platform Containerized solutions (e.g.,
for creating platforms (AWS SageMaker, has components such as MLflow Docker)
machine learn-AzureML, Google AI Platform) Projects and MLflow Model Reg- Serverless computing (AWS
ing pipelines Orchestration platforms istry Lambda, Azure Functions, etc.)
and opera-(Apache Airflow, Jenkins, Kubeflow Continuous
integrationalizing Kubeflow, MLflow, Polyaxon, Kafka-ML tion/continuous deployment
models Seldon Core, Valohai) MLModelCI (CI/CD) tools
Configuration frameworks Monitoring of processes and
(TensorFlow Extended, Gitlab) events
MLOps automation using
AutoML
Containerization for packaging
model dependencies.
        </p>
        <p>API technologies for
deployment as a web service.</p>
        <p>Main features Common features related to Experiment tracking Data, model, and code
versionofered by all phases of machine learning Model packaging and version- ing
MLOps tools pipelines ing Workflow orchestration and
auData management features ML project management tomation
Model management features Model registry Monitoring of processes and
Continuous integration and de- events
livery Integration with cloud and edge
Model monitoring infrastructure
Hyperparameter tuning Containerization
Portability MLOps automation using
AutoML
Methods of Machine learning cloud MLOps is considered as a set of Deployment in the cloud Using
deploying platforms (AWS SageMaker, practices and principles for op- cloud computing resources
machine learn-AzureML, DotScience, Google erationalizing Providing isolation
ing models in AI Platform) Some MLOps tools, such as Hybrid approaches
production Orchestration platforms MLflow, Kubeflow, and Kafka- Deployment on edge/IoT
deenvironments (Apache Airflow, Jenkins, ML vices</p>
        <p>Kubeflow, MLflow, Polyaxon, The role of “Deployment Lead” Deployment directly on IoT
deSeldon Core, Valohai) vices and mobile devices.
TensorFlow Extended (TFX) TensorFlow Lite and Core ML.
Overcoming limitations
Containerized deployments
Packaging models
Docker.</p>
        <p>Serverless architectures
Deploying ML functions as a
service
Reducing costs
Deployment via APIs</p>
        <p>Continued on next page</p>
      </sec>
      <sec id="sec-6-4">
        <title>Continuation of table B.1</title>
        <p>
          Comparison Review by Review by Review by
object Recupito et al. [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] Lima et al. [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] Diaz-de Arcaya et al. [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ]
Maturity mod- Support for continuous inte- Maturity model proposed by The article does not define
speels for assess- gration and continuous deploy-Amershi et al. [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ] cific maturity models for
assessing the level of ment (CI/CD) Dhanorkar et al. [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ] classify or- ing the automation of machine
automation in The ability to automatically ganizations into three levels of learning model deployment, but
deploying ma- tune maturity emphasizes the importance of
chine learning Full automation of model man- Lwakatare et al. [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ] describe adapting software development
models agement processes ifve stages of improvement in practices to the field of machine
development practices learning.
        </p>
        <p>Akkiraju et al. [21] propose an
adaptation of the Capability
Maturity Model (CMM)
Roles and re- Data scientists – responsible for Domain specialist – has deep Software developers
sponsibilities developing and training knowledge of the subject area Data specialists/data scientists
identified in Data engineers – responsible Computational scien- Operations engineers
the activities for extracting, processing, trans- tist/engineer – has high Domain experts
of operational- forming, and ensuring the qual- technical skills Management/stakeholders
izing machine ity of data ML scientist/engineer –
responlearning mod- DevOps engineers – responsi- sible for designing
els ble for automating deployment Provenance specialist –
manprocesses and managing opera- ages the supply of data
tional environments Manager – evaluates models
Product managers and business Application developer –
develstakeholders – provide require- ops applications
ments for models and partici- Deployment lead – assesses
pate in decision-making aspects
Challenges Complexity of managing Integration of software devel- Managing the ML lifecycle
encountered Ensuring consistency opment Gap between software
engineerwhen deploy- Integration of various tools Implementation of ing
ing machine Automation of all stages MLOps/AIOps practices Data management
learning Monitoring model performance Machine learning models need Distributed and parallel
execumodels in Scaling infrastructure monitoring tion
production Ensuring security Identifying infrastructure Diversity of computing
infrasenvironments components tructure</p>
        <p>Deploying and versioning Monitoring</p>
        <p>Explainability
Challenges Lack of standardization Integration of software develop- Lack of skilled personnel
encountered Ensuring portability ment processes Data management issues
when deploy- Configuration and integration Implementation of MLOps prac- Complexity of orchestration
ing machine Full automation tices Heterogeneity of hardware
learning Understandability It is necessary to go beyond an- Need for continuous
monitormodels in alyzing model prototypes ing
production Careful monitoring Lack of explainability
environments Determining infrastructure Scaling and performance issues
Addressing scalability
Need for versioning
Automation
Managing the lifecycle
Integration with DevOps</p>
        <p>Continued on next page</p>
      </sec>
      <sec id="sec-6-5">
        <title>Continuation of table B.1</title>
        <p>
          Comparison Review by Review by Review by
object Recupito et al. [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] Lima et al. [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] Diaz-de Arcaya et al. [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ]
Opportunities, Standardization of MLOps Demand for MLOps tools and Involvement of business units
future trends, practices and tools platforms is expected to grow Greater attention to the ML
and areas of Improved support environ- Industries where MLOps is lifecycle
application of ments. actively developing – finance, Better data management
pracMLOps Advances in automation and healthcare, industry, retail, tices
        </p>
        <p>Integration of MLOps with transportation, and logistics. Use of new hardware platforms
DevOps and DevSecOps. Possible development of spe- Use of containers, serverless
Increased attention to data cialized technologies
management Integration of MLOps with De- Development of versioning
Areas of application vSecOps, MLSecOps concepts tools
Financial services and banking Emergence of new roles and Industries where MLOps is
Healthcare and biotechnology competencies thriving:
Manufacturing and Internet of Traditional industries
Things Innovative industries
Retail and e-commerce Academic disciplines
Telecommunications Communication and
networkTransportation and logistics ing technologies
Healthcare</p>
        <p>Scientific activity</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>P. V.</given-names>
            <surname>Zahorodko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. O.</given-names>
            <surname>Semerikov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. N.</given-names>
            <surname>Soloviev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Striuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. I.</given-names>
            <surname>Striuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. M.</given-names>
            <surname>Shalatska</surname>
          </string-name>
          ,
          <article-title>Comparisons of performance between quantum-enhanced and classical machine learning algorithms on the IBM Quantum Experience</article-title>
          ,
          <source>Journal of Physics: Conference Series</source>
          <year>1840</year>
          (
          <year>2021</year>
          )
          <article-title>012021</article-title>
          . doi:
          <volume>10</volume>
          .1088/
          <fpage>1742</fpage>
          -
          <lpage>6596</lpage>
          /
          <year>1840</year>
          /1/012021.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>D.</given-names>
            <surname>Kreuzberger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Kühl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hirschl</surname>
          </string-name>
          ,
          <article-title>Machine Learning Operations (MLOps): Overview, Definition, and Architecture</article-title>
          ,
          <source>IEEE Access 11</source>
          (
          <year>2023</year>
          )
          <fpage>31866</fpage>
          -
          <lpage>31879</lpage>
          . doi:
          <volume>10</volume>
          .1109/ACCESS.
          <year>2023</year>
          .
          <volume>3262138</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>G.</given-names>
            <surname>Symeonidis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Nerantzis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kazakis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. A.</given-names>
            <surname>Papakostas</surname>
          </string-name>
          , MLOps - Definitions, Tools and Challenges,
          <source>in: 2022 IEEE 12th Annual Computing and Communication Workshop and Conference (CCWC)</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>0453</fpage>
          -
          <lpage>0460</lpage>
          . doi:
          <volume>10</volume>
          .1109/CCWC54503.
          <year>2022</year>
          .
          <volume>9720902</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>M.</given-names>
            <surname>Testi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ballabio</surname>
          </string-name>
          , E. Frontoni, G. Iannello,
          <string-name>
            <given-names>S.</given-names>
            <surname>Moccia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Soda</surname>
          </string-name>
          , G. Vessio,
          <article-title>MLOps: A Taxonomy and a Methodology, IEEE Access 10 (</article-title>
          <year>2022</year>
          )
          <fpage>63606</fpage>
          -
          <lpage>63618</lpage>
          . doi:
          <volume>10</volume>
          .1109/ACCESS.
          <year>2022</year>
          .
          <volume>3181730</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J.</given-names>
            <surname>Diaz-de Arcaya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. I.</given-names>
            <surname>Torre-Bastida</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Zárate</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Miñón</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Almeida</surname>
          </string-name>
          ,
          <string-name>
            <surname>A Joint</surname>
          </string-name>
          <article-title>Study of the Challenges, Opportunities, and Roadmap of MLOps and AIOps: A Systematic Survey</article-title>
          ,
          <source>ACM Comput. Surv</source>
          .
          <volume>56</volume>
          (
          <year>2023</year>
          )
          <article-title>84</article-title>
          . doi:
          <volume>10</volume>
          .1145/3625289.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>G.</given-names>
            <surname>Recupito</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Pecorelli</surname>
          </string-name>
          , G. Catolino,
          <string-name>
            <given-names>S.</given-names>
            <surname>Moreschini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. D.</given-names>
            <surname>Nucci</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Palomba</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. A.</given-names>
            <surname>Tamburri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A Multivocal</given-names>
            <surname>Literature</surname>
          </string-name>
          <article-title>Review of MLOps Tools and Features</article-title>
          ,
          <source>in: 2022 48th Euromicro Conference on Software Engineering and Advanced Applications (SEAA)</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>84</fpage>
          -
          <lpage>91</lpage>
          . doi:
          <volume>10</volume>
          .1109/ SEAA56994.
          <year>2022</year>
          .
          <volume>00021</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>M.</given-names>
            <surname>Steidl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Felderer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ramler</surname>
          </string-name>
          ,
          <article-title>The pipeline for the continuous development of artificial intelligence models-current state of research and practice</article-title>
          ,
          <source>Journal of Systems and Software</source>
          <volume>199</volume>
          (
          <year>2023</year>
          )
          <article-title>111615</article-title>
          . doi:
          <volume>10</volume>
          .1016/j.jss.
          <year>2023</year>
          .
          <volume>111615</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>A.</given-names>
            <surname>Lima</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Monteiro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. P.</given-names>
            <surname>Furtado</surname>
          </string-name>
          , MLOps: Practices, Maturity Models, Roles, Tools, and
          <string-name>
            <surname>Challenges - A Systematic Literature</surname>
            <given-names>Review</given-names>
          </string-name>
          ,
          <source>in: Proceedings of the 24th International Conference on Enterprise Information Systems - Volume 1: ICEIS</source>
          , INSTICC, SciTePress,
          <year>2022</year>
          , pp.
          <fpage>308</fpage>
          -
          <lpage>320</lpage>
          . doi:
          <volume>10</volume>
          .5220/0010997300003179.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>K.</given-names>
            <surname>Haller</surname>
          </string-name>
          ,
          <article-title>Managing AI in the enterprise: Succeeding with AI projects and MLOps to build sustainable AI organizations</article-title>
          , Apress Berkeley, CA,
          <year>2022</year>
          . doi:
          <volume>10</volume>
          .1007/978-1-
          <fpage>4842</fpage>
          -7824-6.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>E. e Oliveira</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Rodrigues</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Pereira</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Lopes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I. I.</given-names>
            <surname>Mestric</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bjelogrlic</surname>
          </string-name>
          ,
          <article-title>Unlabeled learning algorithms and operations: overview and future trends in defense sector</article-title>
          ,
          <source>Artificial Intelligence Review</source>
          <volume>57</volume>
          (
          <year>2024</year>
          )
          <article-title>66</article-title>
          . doi:
          <volume>10</volume>
          .1007/s10462-023-10692-0.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>A. B. Kolltveit</surname>
            ,
            <given-names>J. Li,</given-names>
          </string-name>
          <article-title>Operationalizing machine learning models: a systematic literature review</article-title>
          ,
          <source>in: Proceedings of the 1st Workshop on Software Engineering for Responsible AI</source>
          ,
          <volume>SE4RAI</volume>
          '
          <fpage>22</fpage>
          ,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2023</year>
          , p.
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          . doi:
          <volume>10</volume>
          .1145/3526073. 3527584.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>F.</given-names>
            <surname>Calefato</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Lanubile</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Quaranta</surname>
          </string-name>
          ,
          <article-title>A Preliminary Investigation of MLOps Practices in GitHub</article-title>
          ,
          <source>in: Proceedings of the 16th ACM / IEEE International Symposium on Empirical Software Engineering and Measurement</source>
          , ESEM '22,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2022</year>
          , p.
          <fpage>283</fpage>
          -
          <lpage>288</lpage>
          . doi:
          <volume>10</volume>
          .1145/3544902.3546636.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>M. J. Page</surname>
            ,
            <given-names>J. E.</given-names>
          </string-name>
          <string-name>
            <surname>McKenzie</surname>
            ,
            <given-names>P. M.</given-names>
          </string-name>
          <string-name>
            <surname>Bossuyt</surname>
            , I. Boutron,
            <given-names>T. C.</given-names>
          </string-name>
          <string-name>
            <surname>Hofmann</surname>
            ,
            <given-names>C. D.</given-names>
          </string-name>
          <string-name>
            <surname>Mulrow</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Shamseer</surname>
            ,
            <given-names>J. M.</given-names>
          </string-name>
          <string-name>
            <surname>Tetzlaf</surname>
            ,
            <given-names>E. A.</given-names>
          </string-name>
          <string-name>
            <surname>Akl</surname>
            ,
            <given-names>S. E.</given-names>
          </string-name>
          <string-name>
            <surname>Brennan</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Chou</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Glanville</surname>
            ,
            <given-names>J. M.</given-names>
          </string-name>
          <string-name>
            <surname>Grimshaw</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Hróbjartsson</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. M. Lalu</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>E. W.</given-names>
          </string-name>
          <string-name>
            <surname>Loder</surname>
            , E. Mayo-Wilson,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>McDonald</surname>
            ,
            <given-names>L. A.</given-names>
          </string-name>
          <string-name>
            <surname>McGuinness</surname>
            ,
            <given-names>L. A.</given-names>
          </string-name>
          <string-name>
            <surname>Stewart</surname>
            , J. Thomas,
            <given-names>A. C.</given-names>
          </string-name>
          <string-name>
            <surname>Tricco</surname>
            ,
            <given-names>V. A.</given-names>
          </string-name>
          <string-name>
            <surname>Welch</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Whiting</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Moher</surname>
          </string-name>
          ,
          <article-title>The PRISMA 2020 statement: an updated guideline for reporting systematic reviews</article-title>
          ,
          <source>BMJ</source>
          <volume>372</volume>
          (
          <year>2021</year>
          )
          <article-title>n71</article-title>
          . doi:
          <volume>10</volume>
          .1136/bmj.n71.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>C.</given-names>
            <surname>Haertel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Staegemann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Daase</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Pohl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Nahhas</surname>
          </string-name>
          ,
          <string-name>
            <surname>K. Turowski,</surname>
          </string-name>
          <article-title>MLOps in Data Science Projects: A Review</article-title>
          ,
          <source>in: 2023 IEEE International Conference on Big Data (BigData)</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>2396</fpage>
          -
          <lpage>2404</lpage>
          . doi:
          <volume>10</volume>
          .1109/BigData59044.
          <year>2023</year>
          .
          <volume>10386139</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>R.</given-names>
            <surname>Cohen</surname>
          </string-name>
          , Digital Strategy,
          <source>Machine Learning, and Industry Survey of MLOps, in: Digital Strategies and Organizational Transformation</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>137</fpage>
          -
          <lpage>150</lpage>
          . URL: https://tinyurl.com/33z6zpd3. doi:
          <volume>10</volume>
          . 1142/9789811271984_
          <fpage>0008</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>T. A.</given-names>
            <surname>Sipe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. L.</given-names>
            <surname>Curlette</surname>
          </string-name>
          ,
          <article-title>A meta-synthesis of factors related to educational achievement: a methodological approach to summarizing and synthesizing meta-analyses</article-title>
          ,
          <source>International Journal of Educational Research</source>
          <volume>25</volume>
          (
          <year>1996</year>
          )
          <fpage>583</fpage>
          -
          <lpage>698</lpage>
          . doi:
          <volume>10</volume>
          .1016/S0883-
          <volume>0355</volume>
          (
          <issue>96</issue>
          )
          <fpage>80001</fpage>
          -
          <lpage>2</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>J.</given-names>
            <surname>Chrastina</surname>
          </string-name>
          ,
          <article-title>Meta-synthesis of qualitative studies: background, methodology and applications</article-title>
          ,
          <source>in: NORDSCI Conference proceedings, volume 1 of NORDSCI Conference, Saima Consult Ltd</source>
          ,
          <year>2018</year>
          . doi:
          <volume>10</volume>
          .32008/nordsci2018/b1/v1/13.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>S.</given-names>
            <surname>Amershi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Begel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Bird</surname>
          </string-name>
          , R. DeLine,
          <string-name>
            <given-names>H.</given-names>
            <surname>Gall</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Kamar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Nagappan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Nushi</surname>
          </string-name>
          , T. Zimmermann,
          <source>Software Engineering for Machine Learning: A Case Study, in: 2019 IEEE/ACM 41st International Conference on Software Engineering: Software Engineering in Practice (ICSE-SEIP)</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>291</fpage>
          -
          <lpage>300</lpage>
          . doi:
          <volume>10</volume>
          .1109/ICSE-SEIP.
          <year>2019</year>
          .
          <volume>00042</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>S.</given-names>
            <surname>Dhanorkar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. T.</given-names>
            <surname>Wolf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Qian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Popa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Who needs to know what, when?: Broadening the Explainable AI (XAI) Design Space by Looking at Explanations Across the AI Lifecycle</article-title>
          ,
          <source>in: Proceedings of the 2021 ACM Designing Interactive Systems Conference, DIS '21</source>
          ,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2021</year>
          , p.
          <fpage>1591</fpage>
          -
          <lpage>1602</lpage>
          . doi:
          <volume>10</volume>
          .1145/ 3461778.3462131.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>L. E.</given-names>
            <surname>Lwakatare</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Crnkovic</surname>
          </string-name>
          , J. Bosch,
          <article-title>DevOps for AI - Challenges in Development of AI-enabled Applications</article-title>
          , in: 2020 International Conference on Software, Telecommunications and Computer
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>