<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <article-id pub-id-type="doi">10.18653/v1/2023.findings-emnlp.872</article-id>
      <title-group>
        <article-title>SkeltyMLOps: Orchestrating Collaborative MLOps Activities</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Charbel Daoud</string-name>
          <email>charbel.daoud@mines-ales.fr</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Danielle Azar</string-name>
          <email>danielle.azar@lau.edu.lb</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Julie Boiché</string-name>
          <email>julie.boiche@umontpellier.fr</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Christelle Urtado</string-name>
          <email>christelle.urtado@mines-ales.fr</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sylvain Vauttier</string-name>
          <email>sylvain.vauttier@mines-ales.fr</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science &amp; Mathematics, Lebanese American University</institution>
          ,
          <addr-line>Byblos</addr-line>
          ,
          <country country="LB">Lebanon</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>EuroMov Digital Health in Motion, Univ. Montpellier and IMT Mines Ales</institution>
          ,
          <addr-line>Ales</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>14889</volume>
      <issue>704</issue>
      <fpage>13088</fpage>
      <lpage>13103</lpage>
      <abstract>
        <p>With the increasing adoption of machine learning (ML), the need to manage, in a tailored manner, the complete lifecycle of this new type of software arises. MLOps (Machine Learning Operations) extends DevOps to address the specific challenges of development and lifelong management of ML-powered software. Despite the increasing attention to this domain, practitioners still lack an accessible, modular architecture to support MLOps projects. To address this lack, this paper proposes SkeltyMLOps, a reference architecture designed to promote collaboration between the diverse actors involved in MLOps processes. The architecture is derived from a thorough literature review, from which we extracted and clustered, using a Large Language Model (LLM), a comprehensive list of MLOps actors and activities. These clusters were then used to guide the architectural decomposition and component design of SkeltyMLOps. The originality of our proposed reference architecture is (i) that it clearly aligns its components with the responsibilities of MLOps actors, covering the identified dimensions of MLOps processes, and (ii) mediates collaboration by orchestrating interactions between the diferent types of actors. This architecture provides a foundation for building collaborative ML-powered software.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Software Engineering</kwd>
        <kwd>Machine Learning Operations</kwd>
        <kwd>MLOps Actors</kwd>
        <kwd>MLOps Activities</kwd>
        <kwd>MLOps Reference Architecture</kwd>
        <kwd>Collaborative MLOps</kwd>
        <kwd>Orchestration</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        The advent of machine learning (ML) has led to the adaptation of DevOps practices to manage
ML-powered software, resulting in the rise of Machine Learning Operations (MLOps) which is heavily
inspired by DevOps [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ]. DevOps, a set of established practices within modern software engineering,
integrates development and operational practices to streamline and shorten software release cycles
through systematic build and deployment processes heavily dependent on automation [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Specifically,
the Dev segment in DevOps consists of the Plan, Code, Build, and Test phases, while the Ops segment
consists of the Release, Deploy, Operate, and Monitor phases [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. As software systems evolve, whether
for corrective or evolutionary maintenance, DevOps emphasizes iterative feedback loops that connect
the Monitor phase back to the Plan phase. MLOps emerged around 2015, significantly influenced
by the foundational work of Sculley et al. [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Consequently, the software engineering community
is increasingly addressing ML-specific challenges through MLOps practices. Nevertheless, despite
this increased awareness and adoption, MLOps practitioners continue to face numerous unaddressed
challenges. First, Kolar et al. [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] note the absence of a consolidated MLOps architecture and the
complexity of coordinating multiple technological components. Second, black-box tools with limited
modularity hinder adoption by organizations [
        <xref ref-type="bibr" rid="ref7 ref8">7, 8</xref>
        ]. A preliminary investigation of open-source
contributions on GitHub confirms this adoption barrier, revealing few accessible frameworks or
reference architectures for practitioners[
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Third, collaboration is a critical challenge that afects
the success of ML-powered software which typically relies on multidisciplinary team structures [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ].
However, many MLOps implementations currently lack the tools and practices needed to efectively
facilitate such collaboration. Uysal et al. [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] emphasize the need for a tailored and holistic project
management framework for ML-powered software. Finally, the engineering rigor in ML-powered
software is still maturing, especially in eliciting requirements, which often depends on project-specific
practices and therefore varies in consistency [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. This paper proposes SkeltyMLOps, a reference
architecture intended to facilitate the practical adoption of MLOps practices. SkeltyMLOps is structured
around a set of well-scoped actors, each defined by their responsibilities and interactions, and focuses
on orchestrating the core MLOps activities reported in the literature. The approach for designing
SkeltyMLOps includes three main steps: (1) identifying, analyzing, and mapping the diferent actors
and dimensions of MLOps; (2) identifying the set of activities that constitute an MLOps process; and (3)
clustering these activities into software components to create a reference architecture. In summary,
the contributions of this work are twofold: (a) identification and formalization of the key actors
and activities involved in MLOps; and (b) introduction of SkeltyMLOps, a reference architecture
designed to support the orchestration of these activities and to strengthen the collaboration between actors.
The remainder of this paper is structured as follows. Section 2 outlines the research questions that
guide this work and the methodology used to address them. Section 3 identifies the main MLOps
actors and dimensions, mapping and naming them through a systematic literature review and semantic
clustering via a Large Language Model. Section 4 examines the activities performed by each actor and
the interactions among them within these dimensions. Section 5 discusses threats to the validity and
reliability of the results. Section 6 presents SkeltyMLOps, a reference architecture that orchestrates
these activities and promotes collaboration among actors. Section 7 positions SkeltyMLOps within the
existing literature, and highlights its novel contributions. Finally, Section 8 describes directions for
future work.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Research Questions &amp; Methodology</title>
      <p>This section presents the research questions (RQs) that guide this study, describes the methodology
adopted to address them, and explains the rationale for using a Large Language Model (LLM) within
this research.</p>
      <sec id="sec-2-1">
        <title>2.1. Research Questions</title>
        <p>This research aims to identify the dimensions, actors, and activities of MLOps and to extract the
fundamental components necessary for designing a reference architecture. The formulated RQs that
refine and guide our study are listed below:
∙ RQ1: Who are the main actors involved in MLOps?
∙ RQ2: What are the dimensions of MLOps?</p>
        <p>A dimension is a distinct functional area that groups together related actors and activities.
∙ RQ3: What are the diferent activities that constitute MLOps processes?
∙ RQ4: How should an MLOps architecture that efectively integrates the activities identified in RQ3
be designed to facilitate collaboration among the actors identified in RQ1?</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Methodology for RQ1 and RQ3</title>
        <sec id="sec-2-2-1">
          <title>To address RQ1 and RQ3, a systematic literature review (SLR) is conducted.</title>
          <p>
            Search Strategy. The literature search was carried out via Google Scholar. Using Google Scholar
reduces bias towards specific publishers as suggested by prior comparisons [
            <xref ref-type="bibr" rid="ref13 ref14">13, 14</xref>
            ]. The search
query was initially formulated from key terms appearing in the RQs, including MLOps, actors,
activities, processes, and architecture. To enhance coverage and precision, the query was extended
with relevant synonyms and semantically related terms. The final version of the search string is:
allintitle:(MLOps OR "machine learning operations") AND (actor OR actors OR
pipeline OR pipelines OR architecture OR architectures OR architecting OR
activity OR activities OR workflow OR workflows OR process OR processes).
Selection Criteria. To ensure consistency and reliability in the selection process, two authors
independently reviewed a random sample of 10 papers. A consensus meeting was held to align their
interpretations, resolve ambiguities, and finalize the inclusion and exclusion criteria. It was decided
that a paper will be included if it explicitly presents at least one of the following: (i) a list of MLOps
actors, (ii) a figure describing MLOps processes, or (ii) an enumeration of related activities. A paper will
be excluded if it belongs to: (a) non-English publications, (b) citation-only entries (e.g., bibliographic
listings without content), (c) unpublished manuscripts, (d) inaccessible documents (e.g., full-text
unavailable online or behind paywalls without institutional access), and (e) papers retrieved from
subscription-only sources not accessible to the authors at the time of review.
          </p>
          <p>
            Execution. The Google Scholar search initially1 retrieved 51 entries. After applying the inclusion
and exclusion criteria, the dataset was reduced to 35 primary studies. A random sample of 5
papers was selected from this set to refine the data extraction strategy. Data extraction was
conducted independently by two authors and finalized through a consensus meeting. For RQ1,
the set of actors, explicitly mentioned in each paper, were extracted. For RQ3, the described
activities, either from process diagrams or enumerated descriptions in the text, were extracted.
Following this pilot phase, the remaining papers were each assigned to both authors for
independent extraction. The extracted data were then validated through bilateral discussions to reach consensus.
Snowballing. To complement the initial title-based search and address its limitations, backward
and forward snowballing techniques were applied. Backward snowballing involves reviewing
the reference lists of the initially selected papers to identify additional relevant studies. Forward
snowballing, in turn, identifies studies that cite those papers. Applying the same inclusion and
exclusion criteria, this process yielded 18 additional papers, resulting in a final set of 53 selected
publications. Relevant data to RQ1 and RQ3 were extracted using the same extraction procedures detailed earlier.
Data Synthesis. Out of the 53 retained studies, 38 contribute to answering RQ1 and 45 to RQ3, with
30 papers addressing both. Figure 1 illustrates the review workflow. A data extraction sheet [
            <xref ref-type="bibr" rid="ref15">15</xref>
            ] is
used to systematically collect information from the selected studies. The extracted data consists of: (i)
study metadata such as title, authors, and citation; (ii) for RQ1, a list of explicitly mentioned actors;
and (iii) for RQ3, a list of identified activities, whether extracted from described processes or explicitly
enumerated in the text. The subsequent sections describe how this data was analyzed and mapped to
RQ1 and RQ3.
          </p>
        </sec>
        <sec id="sec-2-2-2">
          <title>1The search was executed in March 2025.</title>
        </sec>
      </sec>
      <sec id="sec-2-3">
        <title>2.3. An LLM-powered Systematic Literature Review</title>
        <p>
          The experiments in this study employ an LLM to systematically cluster and label data extracted
from the systematic literature review. This approach is motivated by recent findings showing that
LLMs can enhance unsupervised clustering by imposing semantic structure on unlabeled data [
          <xref ref-type="bibr" rid="ref16 ref17">16, 17</xref>
          ].
Zhang et al. [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ] propose ClusterLLM, a framework that leverages ChatGPT’s semantic capabilities
to improve clustering outcomes, while maintaining low computational cost. Furthermore, LLMs have
been shown to serve as eficient annotators, capable of producing meaningful semantic labels that
approximate or replace human annotations [
          <xref ref-type="bibr" rid="ref19">19, 20</xref>
          ]. Following a comparative evaluation of multiple
LLMs2, we selected OpenAI’s o3 model3 based on its superior performance in terms of reasoning
quality, completeness (e.g., no missed entries during clustering), and labeling consistency [21, 22]. The
prompts used throughout the clustering process were iteratively refined according to established prompt
engineering practices [23, 24].
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. MLOps Actors &amp; Dimensions</title>
      <p>This section jointly addresses RQ1 and RQ2 by identifying actors and dimensions involved in MLOps.
The analysis follows an iterative process in which the dimensions are inferred from the initially identified
actors. These dimensions are then used to re-examine the actors, allowing for the refinement of the
actor set. This process ensures that both RQs cross-validate each other.</p>
      <sec id="sec-3-1">
        <title>3.1. MLOps Actors</title>
        <p>The actors identified when addressing RQ1 (Section 2.2) are analyzed. Table 1 summarizes actors
referenced in multiple sources along with their corresponding references. Actors mentioned only in a
single study are excluded from the table (such as Data Provider [38] or ML Owner [56]), as their limited
appearance suggests they are not widely recognized. During the extraction process, inconsistencies
in how actors and their responsibilities were defined across studies were noted. For example, while
most papers assign Data Scientists to data-centric tasks, others also associate them with model
development responsibilities. Similarly, the role of Software Engineers varies significantly: some
sources limit their involvement in software development, whereas others include operational and ML
tasks. To systematically reduce such ambiguities and identify actor groups, we apply a clustering
approach based on the LLM selected in Section 2.3. This approach is described in the next subsection.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. MLOps Dimensions</title>
        <p>To address the ambiguities stemming from unclear actor responsibilities, dimensions are formally
identified and used to structure actors. We define an MLOps dimension as a conceptual axis that
characterizes a specific concern in MLOps. Together, these dimensions form a conceptual space in
which actors can be situated based on their responsibilities along each axis. This representation
facilitates a structured understanding of collaboration across MLOps. This process is carried out using
the o3-model through a three-step process: (1) identify relevant dimensions from a predefined actor set,
(2) cluster actors according to these dimensions, and (3) propose representative actor name for each
cluster. The exact prompt provided to the o3 model is included below:
You’re an expert in machine learning operations (MLOps). You will find below a list that contains 16 Actors who
appear in diferent activities in an MLOps process. I want you to:
• Identify dimensions from the provided actors.
• Cluster actors according to these dimensions.
• Propose a leading actor for each dimension.
• Format the final output as a structured table with three columns: Dimension, List of Actors, Proposed</p>
        <p>Leading Actor.</p>
        <sec id="sec-3-2-1">
          <title>2See the replication package for full details of the LLM comparison. 3OpenAI’s o3 model (see link: https://openai.com/index/introducing-o3-and-o4-mini/)</title>
          <p>
            A manual review is conducted on the raw output generated by the prompt. One adjustment involves renaming
the first dimension from Business &amp; Domain to Plan, to align it with the standard DevOps phases (Plan, Code,
Build, etc.). The actor clusters were also reviewed and updated: the leading actor in the Plan dimension was
changed from Business Lead to Product Owner to reflect the inclusion of software engineering activities
alongside business concerns. Importantly, the Product Owner actively collaborates, within this dimension,
with all other identified actors ensuring alignment and collaboration during the planning phase. All manual
modifications are documented in the replication package [
            <xref ref-type="bibr" rid="ref15">15</xref>
            ]. The final adjusted set of dimensions and leading
actors is presented in Table 2.
          </p>
          <p>Highlights (RQ1 &amp; RQ2). In response to RQ1, we identify 5 actors that lead MLOps dimensions: Product
Owner, Data Engineer, Model Engineer, Software Engineer, and Operations Engineer. Addressing RQ2, we
propose 5 MLOps dimensions: Plan, Data Engineering (Data), Model Engineering (Model), Software
Engineering (Soft), and Operations (Ops).</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. MLOps Activities</title>
      <p>This section further addresses RQ3, investigating how MLOps activities can be systematically identified and
structured. Given that MLOps has no widely accepted standardized set of activities, a six-step methodology is
proposed to fill this gap:
- Activity Extraction. The extraction process, described in Section 2.2, produces a matrix with 402 activities
(rows) across 45 publications (columns).
- Grouping by Dimensions. The 402 recorded activities are then clustered according to the 5 dimensions
introduced in Section 3.2 (Plan, Data, Model, Soft, and Ops ). To automate this mapping, the o3 model is reused.
The complete list of activities is fed into the model, leveraging its language understanding capabilities to classify
each activity under one of the 5 dimensions. The prompt used is:
You’re an expert in machine learning operations (MLOps). You will find, at the end, a list that contains 402
Activities that appear in diferent MLOps processes. I want you to:
• Categorize these Activities into these 5 dimensions: Plan, Data engineering, Model engineering, Software
engineering, and Operations.
• Ensure activities are not mistakenly categorized: e.g., Diferentiate what’s Software and what’s Operations.
• Format the final output as a structured table with two columns: Dimension, Activity. I should get 402 rows
(other than the header)
- Dimension-Specific Clustering. After executing the prompt, the o3 model produces clusters for the Plan (26
activities), Data (101 activities), Model (113 activities), Soft (59 activities), and Ops (103 activities) dimensions.
- Redundancy Reduction. To refine the activity sets extracted per dimension, a dedicated prompt is applied to
each one. This step aims to: (1) group semantically similar activities and eliminate redundancies, and (2) assign a
representative label and generate a concise description for each resulting cluster. The procedure is repeated
independently for each of the 5 dimensions. The prompt used in this step is as follows:
You’re an expert in machine learning operations (MLOps). You will find, at the end, a list of activities related to
&lt;dimension_name&gt; Dimension. You are asked to:
• Group similar activities and propose new activity name. The proposed activity name should start with a
verb. Use concise verb phrase, for example I prefer "Review Code" instead of "perform code review".
• Ensure activities are not mistakenly merged if they represent distinct tasks
• Format the final output as a structured table with three columns: Proposed Activity Name, Grouped</p>
      <p>Activities, A brief explanation. Order the activities in the table in a chronological order.</p>
      <p>This step reduces the total number of distinct activities in Plan, Data, Model, Soft , and Ops dimensions to 7, 8, 9,
5, and 9, respectively (38 activities in total).
- Manual Review and Merging. A manual review of these 38 activities is conducted to refine the activity set
and reduce redundancy4. For example, Diagram Deployment Model was merged with Design Architecture.
- Final Consolidation. The consolidated list of activities, each with a description and frequency across the 45
analyzed publications, is included in the replication package. A visual summary of these activities is provided in
Figure 2, which organizes them according to the 5 proposed MLOps dimensions. Each dimension is color-coded
and annotated with the lead actors responsible for executing their associated activities.</p>
      <p>Highlights (RQ3). In response to RQ3, 38 activities are identified, distributed across the five MLOps
dimensions. These activities constitute the foundation of SkeltyMLOps ’s architectural design.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Threats to Validity</title>
      <p>Despite our methodical approach and the rigor with which the extraction, processing, and analysis of data were
handled, there are still potential threats to validity regarding the obtained results.</p>
      <p>External validity. Our study is based on a selection of academic research papers. This may omit relevant
alternative sources, such as white papers and online articles published by practitioners. However, it is a deliberate
4The details of this manual refinement are provided in the replication package: https://anonymous.4open.science/r/mlops_
ecai25-EB68/
design choice to only retain peer-reviewed papers, corresponding to established practices. Moreover, these
papers are retrieved from a single search engine. We mitigate this threat by relying on Google Scholar, which is
reputed for its agnostic and wide-ranging indexing of publications, and by applying backward and forward
snowballing techniques to cover as many relevant studies as possible.</p>
      <p>Internal validity. The proposed lists of actors, dimensions, and activities may reflect biases introduced by our
backgrounds in software engineering, machine learning, and operations. To reduce these biases, we used an
LLM to group the extracted information. While this approach minimizes subjective manual interpretations, it
introduces new risks inherent to LLMs, such as hallucinations. We partially mitigate this by manually reviewing
and refining the model’s output to ensure coherence and accuracy.</p>
      <p>Reliability. Our results depend on the choice of the LLM and are influenced by the design of the prompts. This
is especially true for the activity clustering task, which involves a large and diverse set of items. LLM responses
can vary across runs or model versions and may lack reproducibility if prompt designs are not well-documented.
Furthermore, since LLMs are trained on vast and heterogeneous data, they might introduce contextually irrelevant
groupings. To address these issues, we iteratively refined the prompts and ensured that outputs were reproducible
under stable conditions. Our final manual adjustments were minimal and limited to coherence and clarity, not
conceptual alterations.</p>
    </sec>
    <sec id="sec-6">
      <title>6. A Reference MLOps Architecture Promoting Collaboration</title>
      <p>This section introduces SkeltyMLOps, a reference architecture for MLOps, that explicitly supports the actors
and activities identified in Sections 3 and 4. A Reference Architecture is a reusable blueprint that encapsulates
essential design principles, best practices, and structural guidelines for a given domain [25, 26]. It serves as a
foundation for developing concrete MLOps frameworks. SkeltyMLOps is a reference architecture for MLOps
that integrates relevant domain-specific knowledge, and standardizes activities within the MLOps process. The
motivation behind SkeltyMLOps is to emphasize and orchestrate collaboration among the various actors involved.
Preprocess Data, Validate Data, Engineer Features, Label Data, Analyze Data, Manage Data
Run Experiments, Design Model, Develop Model, Train Model, Tune Model, Validate Model,
Evaluate &amp; Test Model, Optimize Model, Register Model
Develop Code, Build Components, Package Artifacts, Test Solution, Release Solution
Configure CI/CD, Manage Infrastructure, Deploy to Testing, Deploy to Production, Serve
Inference, Monitor Performance, Operate &amp; Maintain</p>
      <p>Execute Pipelines, Trigger Retraining</p>
      <sec id="sec-6-1">
        <title>6.1. From Actors &amp; Dimensions to an MLOps Reference Architecture</title>
        <p>To design the architecture of SkeltyMLOps, a modular decomposition is adopted, aligned with MLOps identified
dimensions. Each module corresponds to one of the 5 MLOps dimensions and is structured to serve the needs of
its actors. This design follows a domain-driven approach, where each module represents a distinct sub-domain of
MLOps. Within each module, finer-grained components are defined to support the execution of specific activities
identified in Section 4. The overall architecture, with its modules and their internal components, is presented
in Figure 3, while Table 3 details the correspondence between components and the activities they support. At
the core of the architecture is the MLOps Process Orchestration (MPO) component. It manages both control
and data flows among all components, and ensures coordinated execution of the MLOps process. The Project
Management component, which handles requirements engineering and planning tasks, supplies the MPO with
issue tickets [27, 28]. These tickets are then resolved automatically via a translation into executable commands
that involve the appropriate components. A key aspect of SkeltyMLOps is that it versions not only code, but data
and ML models as well. This versioning capability supports reproducibility, allows for rollback when needed,
and can be used to support audits and explanations. Moreover, the architecture separates two complementary
components: API Development and Deployment &amp; Monitoring. The API Development component focuses on
implementing and exposing model functionalities. In parallel, the Deployment &amp; Monitoring component ensures
that these models are monitored and maintained in production environments. This separation clarifies the
boundary between development and operations. The architecture further incorporates components to manage
Data and ML models. The Data Acquisition component provides interfaces for ingesting data from diverse
sources. The Data Processing &amp; Feature Engineering component manages the cleaning and transformations of
raw data, and performs feature engineering. The Model Development &amp; Training component groups services
related to building, training, validating, and optimizing models. The proposed reference architecture allows
updates and recovery procedures to have a minimal impact on other components. To support software evolution,
SkeltyMLOps incorporates feedback loops at multiple levels. The Deployment &amp; Monitoring component enables
continuous monitoring and rollback mechanisms. In addition, it reports operational insights, such as data and
concept drift, back to Project Management, where they are tracked as issue tickets. This feedback loop triggers
maintenance actions within the MLOps process. In addition, end-user feedback collected through the Project
Management interface is systematically integrated into the iterative development process. The Data Acquisition
component further enhances adaptability by exposing interfaces for ingesting new data, thereby supporting
feedback from production, enabling continuous training, and facilitating adaptation to evolving datasets. Finally,
SkeltyMLOps ofers client-facing interfaces that support integration with external tools, including IDEs. The
modular nature of SkeltyMLOps makes it extensible and customizable. Components can be individually adapted,
replaced, or omitted according to project-specific needs such as swapping the Model Storage implementation or
disabling API Development in batch-only pipelines. This modularity enhances the architecture’s maintainability,
reusability, and adaptability.</p>
      </sec>
      <sec id="sec-6-2">
        <title>6.2. Emphasis on MLOps Project Management &amp; Process Orchestration</title>
        <p>The core contribution of SkeltyMLOps lies in its explicit support for actor collaboration and activity orchestration
across the MLOps processes. Unlike existing approaches that treat MLOps as a loosely connected set of
tools and practices, the proposed design introduces structured components that enforce coordination and
operational continuity between dimensions. At the core of this architecture are two central components: Project
Management and MLOps Process Orchestration. The Project Management component initiates MLOps activities.
It ingests BPMN processes and feeds them into the orchestration layer to automate process execution. The
Product Owner, acting as a liaison across all actors, defines the MLOps processes by capturing requirements,
aligning them with business goals, formulating the problem statement, and ensuring communication among
actors. The Product Owner also supervises the MLOps backlog and issue tickets, ensuring alignment between
evolving project needs and technical execution. Complementing this, the MPO component operationalizes the
strategic decisions defined during project planning. Functioning as an execution layer, it continuously monitors
process execution, handles feedback from runtime operations, and adapts system behavior when necessary.
Through this mechanism, the MPO enforces ongoing coordination, ensuring that collaborative intentions are
reflected in actual system behavior. Together, these components form the collaborative and orchestration
backbone of the presented reference architecture. Their multicolored representation in Figure 3 specifically
highlights the interaction and cooperation among the various actors involved within these two components.</p>
        <p>Highlights (RQ4). SkeltyMLOps is an MLOps reference architecture designed to promote collaboration
through the orchestration of activities. These activities are extracted from the literature, associated with
identified MLOps actors from the literature review, and organized into components. The components are
grouped into modules that reflect MLOps dimensions.</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>7. State of the Art on Reference Architectures for MLOps</title>
      <p>
        This section examines recent MLOps reference architectures to contextualize SkeltyMLOps within the broader
literature. Wozniak et al. [29] propose an MLOps reference architecture derived from a systematic literature review,
identifying key components, tools, processes, and metrics. Although their proposed process explicitly integrates
data-related activities, the final presented architecture itself lacks clearly defined data management components.
This omission contrasts with our previous findings, which emphasize the importance of data management as
critical to MLOps processes, a gap explicitly addressed by SkeltyMLOps. In contrast, Najafabadi et al. [30] conduct
a systematic mapping study, identifying 35 architectural components. While their work shares motivations
with SkeltyMLOps, it lists components without defining clear responsibilities or actors among them, causing
overlap and potential confusion. Although Najafabadi et al. highlight the necessity of an orchestrator to manage
interactions among components, their final architecture does not explicitly depict or detail this orchestrator. In
contrast, SkeltyMLOps explicitly defines actors and their interactions based on collaboration, clearly integrating
a central orchestrator into the architecture to coordinate the activities. Kreuzberger et al. [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] present an
end-toend MLOps architecture detailing various components and associated actors. However, their data versioning
strategy primarily focuses on features while neglecting raw data and training/test dataset versioning, which
are critical for reproducibility and traceability in MLOps. Additionally, their monitoring primarily covers model
performance without addressing data drift, resource utilization, or overall infrastructure health, potentially
hindering proactive issue detection and automated retraining processes. SkeltyMLOps addresses these limitations
by providing explicit data versioning strategies across all stages of data lifecycle management and implementing
comprehensive monitoring to proactively manage and adapt to performance, dataset, and model changes. In
SkeltyMLOps, components are systematically organized by actors to facilitate clear and efective collaboration.
At the core of SkeltyMLOps is a central orchestrator designed to coordinate interactions and ensure eficient
execution across components. Additionally, SkeltyMLOps explicitly includes structured monitoring with clearly
defined feedback loops, enabling systematic initiation of retraining or adaptation processes.
      </p>
    </sec>
    <sec id="sec-8">
      <title>8. Conclusion &amp; Perspectives</title>
      <p>Recognizing the specific challenges posed by development and lifelong management of ML-powered software, we
introduce SkeltyMLOps, a reference architecture tailored to address these challenges. The design of SkeltyMLOps
follows a structured methodology: first, identifying and mapping the diverse actors and dimensions involved in
MLOps, second, defining the core set of activities necessary for an MLOps process, and third, organizing these
activities into software components assembled in a reference architecture. Compared to the existing literature,
SkeltyMLOps stands out by focusing on actors and promoting a holistic project management and collaboration
through its architecture design. We envision several extensions to this work. First, we aim to provide the LLM
with the papers from the literature review as context to facilitate more precise reasoning. Second, we will
design generic collaborative MLOps processes and their variants using BPMN. Third, we plan to implement a
BPMN-based orchestrator for SkeltyMLOps to automate and coordinate the modeled MLOps processes. Fourth,
we have initiated the development of an open-source version of SkeltyMLOps. This framework will integrate
widely adopted tools and libraries. Finally, we intend to evaluate SkeltyMLOps in continuous training scenarios.
Acknowledgments. This research was partially supported by Région Occitanie (PLANETS project), which funds
Charbel Daoud’s PhD, and by the French National Research Agency (ANR) (AdaptiveMLOps, ANR-24-IAS2-0004),
which supports the work of Christelle Urtado and Sylvain Vauttier.</p>
    </sec>
    <sec id="sec-9">
      <title>Declaration on Generative AI</title>
      <p>As discussed in Section 2.3, a large language model was used to cluster data extracted from the systematic
literature review.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>D.</given-names>
            <surname>Kreuzberger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Kühl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hirschl</surname>
          </string-name>
          ,
          <article-title>Machine Learning Operations (MLOps): Overview, Definition, and Architecture</article-title>
          ,
          <source>IEEE Access 11</source>
          (
          <year>2023</year>
          )
          <fpage>31866</fpage>
          -
          <lpage>31879</lpage>
          . doi:
          <volume>10</volume>
          .1109/ACCESS.
          <year>2023</year>
          .
          <volume>3262138</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>D. A.</given-names>
            <surname>Tamburri</surname>
          </string-name>
          , Sustainable MLOps:
          <article-title>Trends and Challenges, in: The 22nd International Symposium on Symbolic and Numeric Algorithms for Scientific Computing (SYNASC</article-title>
          , IEEE, Online,
          <year>2020</year>
          , pp.
          <fpage>17</fpage>
          -
          <lpage>23</lpage>
          . doi:
          <volume>10</volume>
          .1109/SYNASC51798.
          <year>2020</year>
          .
          <volume>00015</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>L.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Bass</surname>
          </string-name>
          ,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <article-title>Champlin-Scharf, DevOps</article-title>
          and Its Practices,
          <source>IEEE Software 33</source>
          (
          <year>2016</year>
          )
          <fpage>32</fpage>
          -
          <lpage>34</lpage>
          . doi:
          <volume>10</volume>
          . 1109/MS.
          <year>2016</year>
          .
          <volume>81</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>B. S.</given-names>
            <surname>Farroha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. L.</given-names>
            <surname>Farroha</surname>
          </string-name>
          ,
          <article-title>A Framework for Managing Mission Needs, Compliance, and Trust in the DevOps Environment</article-title>
          , in: IEEE Military Communications Conference (MILCOM), IEEE, Baltimore,
          <string-name>
            <surname>MD</surname>
          </string-name>
          , USA,
          <year>2014</year>
          , pp.
          <fpage>288</fpage>
          -
          <lpage>293</lpage>
          . doi:
          <volume>10</volume>
          .1109/MILCOM.
          <year>2014</year>
          .
          <volume>54</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>D.</given-names>
            <surname>Sculley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Holt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Golovin</surname>
          </string-name>
          , E. Davydov,
          <string-name>
            <given-names>T.</given-names>
            <surname>Phillips</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ebner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Chaudhary</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Young</surname>
          </string-name>
          ,
          <article-title>Machine Learning: The High-Interest Credit Card of Technical Debt</article-title>
          , in: SE4ML:
          <article-title>Software Engineering for Machine Learning</article-title>
          (NeurIPS 2014 Workshop), Montréal,
          <string-name>
            <surname>QC</surname>
          </string-name>
          , Canada,
          <year>2014</year>
          . URL: https://research.google/pubs/ machine
          <article-title>-learning-the-high-interest-credit-card-of-technical-debt/.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A. K.</given-names>
            <surname>Narayanappa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Amrit</surname>
          </string-name>
          ,
          <article-title>An Analysis of the Barriers Preventing the Implementation of MLOps</article-title>
          , in: S. K. Sharma,
          <string-name>
            <given-names>Y. K.</given-names>
            <surname>Dwivedi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Metri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Lal</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          . Elbanna (Eds.),
          <article-title>Transfer, Difusion and Adoption of Next-Generation Digital Technologies</article-title>
          , volume
          <volume>697</volume>
          <source>of IFIP Advances in Information and Communication Technology</source>
          , Springer, Cham,
          <year>2024</year>
          , pp.
          <fpage>101</fpage>
          -
          <lpage>114</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>031</fpage>
          -50188-3_
          <fpage>10</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Burgueño-Romero</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Benítez-Hidalgo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Barba-González</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. F.</given-names>
            <surname>Aldana-Montes</surname>
          </string-name>
          ,
          <article-title>Toward an Open Source MLOps Architecture</article-title>
          ,
          <source>IEEE Software 42</source>
          (
          <year>2025</year>
          )
          <fpage>59</fpage>
          -
          <lpage>64</lpage>
          . doi:
          <volume>10</volume>
          .1109/MS.
          <year>2024</year>
          .
          <volume>3421675</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>A.</given-names>
            <surname>Melde</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Madan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Gavrikov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Hoof</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Laubenheimer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Keuper</surname>
          </string-name>
          , C. Reich,
          <article-title>Tackling Key Challenges of AI Development-Insights from an Industry-Academia Collaboration</article-title>
          , in: C. Reich, U. Mescheder (Eds.),
          <source>The Upper-Rhine Artificial Intelligence Symposium</source>
          (UR-AI
          <year>2022</year>
          )
          <article-title>: AI Applications in Medicine and Manufacturing</article-title>
          , Furtwangen University, Villingen-Schwenningen, Germany,
          <year>2022</year>
          , pp.
          <fpage>112</fpage>
          -
          <lpage>121</lpage>
          . URL: https://opus.hs-furtwangen.de/files/8624/Tackling.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>F.</given-names>
            <surname>Calefato</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Lanubile</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Quaranta</surname>
          </string-name>
          ,
          <article-title>A Preliminary Investigation of MLOps Practices in GitHub</article-title>
          ,
          <source>in: 16th ACM/IEEE International Symposium on Empirical Software Engineering and Measurement (ESEM</source>
          , ACM, Helsinki, Finland,
          <year>2022</year>
          , pp.
          <fpage>283</fpage>
          -
          <lpage>288</lpage>
          . doi:
          <volume>10</volume>
          .1145/3544902.3546636.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>A.</given-names>
            <surname>Serban</surname>
          </string-name>
          , K. van der Blom, H. Hoos,
          <string-name>
            <given-names>J.</given-names>
            <surname>Visser</surname>
          </string-name>
          ,
          <article-title>Software Engineering Practices for Machine Learning-Adoption, Efects, and Team Assessment</article-title>
          ,
          <source>Journal of Systems and Software</source>
          <volume>209</volume>
          (
          <year>2024</year>
          )
          <article-title>111907</article-title>
          . doi:
          <volume>10</volume>
          .1016/j.jss.
          <year>2023</year>
          .
          <volume>111907</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>M. P.</given-names>
            <surname>Uysal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Akturk</surname>
          </string-name>
          ,
          <article-title>A Systemic Approach to Machine Learning Project Management</article-title>
          ,
          <source>IEEE Engineering Management Review</source>
          <volume>52</volume>
          (
          <year>2024</year>
          )
          <fpage>1</fpage>
          -
          <lpage>17</lpage>
          . doi:
          <volume>10</volume>
          .1109/EMR.
          <year>2024</year>
          .
          <volume>3503677</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>M.</given-names>
            <surname>Zarour</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Alzabut</surname>
          </string-name>
          , K. T. Al-Sarayreh, MLOps Best Practices, Challenges and
          <string-name>
            <given-names>Maturity</given-names>
            <surname>Models</surname>
          </string-name>
          :
          <string-name>
            <given-names>A Systematic</given-names>
            <surname>Literature</surname>
          </string-name>
          <string-name>
            <surname>Review</surname>
          </string-name>
          ,
          <source>Information and Software Technology</source>
          <volume>183</volume>
          (
          <year>2025</year>
          )
          <article-title>107733</article-title>
          . doi:
          <volume>10</volume>
          .1016/j. infsof.
          <year>2025</year>
          .
          <volume>107733</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>A.</given-names>
            <surname>Ampatzoglou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bibi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Avgeriou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Verbeek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Chatzigeorgiou</surname>
          </string-name>
          , Identifying, Categorizing and Mitigating Threats to Validity in Software Engineering Secondary Studies,
          <source>Information and Software Technology</source>
          <volume>106</volume>
          (
          <year>2019</year>
          )
          <fpage>201</fpage>
          -
          <lpage>230</lpage>
          . doi:
          <volume>10</volume>
          .1016/j.infsof.
          <year>2018</year>
          .
          <volume>10</volume>
          .006.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>A.</given-names>
            <surname>Martín-Martín</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Orduña-Malea</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Thelwall</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. D.</given-names>
            <surname>López-Cózar</surname>
          </string-name>
          ,
          <article-title>Google Scholar, Web of Science, and Scopus: A systematic comparison of citations in 252 subject categories</article-title>
          ,
          <source>Journal of Informetrics</source>
          <volume>12</volume>
          (
          <year>2018</year>
          )
          <fpage>1160</fpage>
          -
          <lpage>1177</lpage>
          . doi:
          <volume>10</volume>
          .1016/j.joi.
          <year>2018</year>
          .
          <volume>09</volume>
          .002.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>C.</given-names>
            <surname>Daoud</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Azar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Boiché</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Urtado</surname>
          </string-name>
          , S. Vauttier, SkeltyMLOps: Orchestrating Collaborative MLOps Activities, Zenodo, Version
          <volume>0</volume>
          .1,
          <year>2025</year>
          . doi:
          <volume>10</volume>
          .5281/zenodo.17094711, [Dataset].
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>P.</given-names>
            <surname>Trivedi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Choudhary</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. W.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. N.</given-names>
            <surname>Ioannidis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Subbian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Koutra</surname>
          </string-name>
          ,
          <article-title>Large Language Model Guided Graph Clustering</article-title>
          ,
          <source>in: Learning on Graphs Conference (LoG)</source>
          ,
          <source>Virtual Event</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>V.</given-names>
            <surname>Viswanathan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Gashteovski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Lawrence</surname>
          </string-name>
          , T. Wu, G. Neubig,
          <string-name>
            <surname>Large Language Models Enable Few-Shot</surname>
            <given-names>Clustering</given-names>
          </string-name>
          ,
          <source>Transactions of the Association for Computational Linguistics</source>
          <volume>12</volume>
          (
          <year>2024</year>
          )
          <fpage>321</fpage>
          -
          <lpage>333</lpage>
          . doi:
          <volume>10</volume>
          .1162/ tacl_a_
          <fpage>00648</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. Shang,</surname>
          </string-name>
          <article-title>ClusterLLM: Large Language Models as a Guide for Text Clustering, in: Empirical Methods in Natural Language Processing (EMNLP), Association for Computational Linguistics</article-title>
          , Singapore,
          <year>2023</year>
          , pp.
          <fpage>13903</fpage>
          -
          <lpage>13920</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2023</year>
          .emnlp-main.
          <volume>858</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>M. J. Mirza</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Karlinsky</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Possegger</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Kozinski</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Feris</surname>
          </string-name>
          , H. Bischof,
          <article-title>LaFTer: Label-Free tuning of Zero-shot classifier using language and unlabeled image collections</article-title>
          , in: A.
          <string-name>
            <surname>Oh</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Naumann</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Globerson</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Saenko</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Hardt</surname>
          </string-name>
          , S. Levine (Eds.),
          <source>Advances in Neural Information Processing Systems</source>
          <volume>36</volume>
          (
          <issue>NeurIPS</issue>
          ), volume
          <volume>36</volume>
          ,
          <string-name>
            <surname>Curran</surname>
            <given-names>Associates</given-names>
          </string-name>
          , Inc., New Orleans, LA, USA,
          <year>2023</year>
          , pp.
          <fpage>5765</fpage>
          -
          <lpage>5777</lpage>
          . URL: https://proceedings.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>