<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>ML, October</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>TAIBOM: Bringing Trustworthiness to AI-Enabled Systems</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Vadim Safronov</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Anthony McCaigue</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nicholas Allott</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andrew Martin</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>NquiringMinds</institution>
          ,
          <country country="UK">UK</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>University of Oxford</institution>
          ,
          <country country="UK">UK</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>26</volume>
      <issue>2025</issue>
      <fpage>0000</fpage>
      <lpage>0001</lpage>
      <abstract>
        <p>The growing integration of open-source software and AI-driven technologies has introduced new layers of complexity into the software supply chain, challenging existing methods for dependency management and system assurance. While Software Bills of Materials (SBOMs) have become critical for enhancing transparency and traceability, current frameworks fall short in capturing the unique characteristics of AI systems - namely, their dynamic, data-driven nature and the loosely coupled dependencies across datasets, models, and software components. These challenges are compounded by fragmented governance structures and the lack of robust tools for ensuring integrity, trust, and compliance in AI-enabled environments. In this paper, we introduce Trusted AI Bill of Materials (TAIBOM) - a novel framework extending SBOM principles to the AI domain. TAIBOM provides (i) a structured dependency model tailored for AI components, (ii) mechanisms for propagating integrity statements across heterogeneous AI pipelines, and (iii) a trust attestation process for verifying component provenance. We demonstrate how TAIBOM supports assurance, security, and compliance across AI workflows, highlighting its advantages over existing standards such as SPDX and CycloneDX. This work lays the foundation for trustworthy and verifiable AI systems through structured software transparency.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        The rapid expansion of open-source software and AI-driven technologies has introduced unprecedented
complexity into the software supply chain. Managing dependencies is already a challenging task, but
the integration of AI with digital infrastructures further amplifies the dificulty [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ]. In response,
Software Bills of Materials (SBOMs) have emerged as a mechanism for improving accountability and
traceability across software ecosystems [
        <xref ref-type="bibr" rid="ref3 ref4">3, 4</xref>
        ].
      </p>
      <p>
        While various organisations, researchers, and developers acknowledge the growing importance and
usefulness of SBOMs, the absence of mature tools for their generation and application in assurance,
security, and compliance remains a significant gap [
        <xref ref-type="bibr" rid="ref1 ref5 ref6">1, 5, 6</xref>
        ]. This challenge is particularly significant
in AI-enabled systems, where traditional SBOM frameworks struggle to accommodate AI-specific
complexities. AI models are inherently dynamic and data-driven, continuously evolving through
updates and retraining. Ensuring stability and version control is a persistent challenge. Furthermore,
AI systems comprise multiple interdependent components — training datasets, training software,
refinement datasets, trained weights, inference software — that are often loosely coupled, making
dependency tracking and provenance verification dificult. The governance of AI systems is also
complicated by their distributed nature, where data ownership, model training, and system deployment
often reside with diferent entities, making the enforcement of security, privacy, and safety standards
particularly challenging.
      </p>
      <p>
        Although there is much interest in Trust in AI [
        <xref ref-type="bibr" rid="ref7 ref8">7, 8, 9</xref>
        ], much of the discussion is predicated on the
accurate identification of (and so tamer-proofing of) all these diverse aforementioned elements, which
in turn raises several challenging questions. Can the integrity of model producers be verified? Can
datasets — often assembled from multiple sources — be trusted? Can transparency and accountability
be ensured throughout the AI development pipeline?
      </p>
      <p>To address these challenges, we introduce the Trusted AI Bill of Materials (TAIBOM) — an evolution
of SBOM frameworks extended to AI-enabled software systems. TAIBOM is accompanied by an
implementation toolkit 1, designed to support its principles. The key contributions of TAIBOM are:
• A structured dependency model that conceptualises AI software components and their
interrelationships.
• A framework for propagating integrity statements across disconnected AI environments (e.g.,
dataset, training, and inference), ensuring a continuous chain of trust.</p>
      <p>• A trust attestation mechanism to verify the provenance and trustworthiness of AI components.</p>
      <p>This paper is organised as follows. Section 2 overviews SBOMs and motivates the need for AIBOMs.
Section 3 reviews related work, identifies limitations in existing AIBOM standards regarding trust
guarantees, and outlines how TAIBOM addresses them. Section 4 presents the TAIBOM dependency
model and its component relationships. Section 5 demonstrates TAIBOM’s integration into AI workflows
across assurance, security, compliance, and risk management use cases, and compares its functionality
to state-of-the-art AIBOM approaches. Section 6 discusses current limitations and future directions.
Section 7 concludes the paper.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Background</title>
      <p>To introduce the need for trust in AI-enabled systems, this background section reviews the concept
of Software Bills of Materials (SBOMs), their role in traditional software supply chains, and how they
inform the emerging notion of AI-specific Bills of Materials (AIBOMs).</p>
      <sec id="sec-2-1">
        <title>2.1. Software Bill of Materials</title>
        <p>The Software Bill of Materials (SBOM), inspired by manufacturing industry practices, was formalised
in 2018 by the US National Telecommunications and Information Administration (NTIA) to enhance
software security practices and has since evolved [10]. An SBOM provides a detailed inventory of
software components, specifying their origins, dependencies, and references to known or potential
vulnerabilities. Key elements in the SBOM ecosystem include the Common Platform Enumeration (CPE,
a standardised naming scheme for software components [11]), Common Vulnerabilities and Exposures
(CVE, a repository of publicly disclosed cybersecurity vulnerabilities [12]), and Common Weakness
Enumeration (CWE, a classification system for software weaknesses [13]).</p>
        <p>SBOMs ensure that all software components are identified and traceable. However, concerns remain
about their granularity. Various SBOM generation approaches have emerged. Binary-focused tools [14,
15, 16] analyse compiled binaries using embedded metadata, string literals, and language-specific
features to identify dependencies. Metadata-based tools [17, 18, 19] extract dependency information
from package metadata, build files, and container images. Source code analysis tools [ 20, 21, 22] inspect
repositories to uncover dependencies, including hidden ones associated with known CVEs.</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. AI Bill of Materials</title>
        <p>
          While existing SBOM approaches are suitable for traditional software, they fall short in addressing
vulnerabilities specific to AI-enabled systems. Some argue that AI is merely another category of software
and can be managed under current SBOM regulations. However, recent research suggests otherwise,
highlighting a growing demand for an AI Bill of Materials (AIBOM) capable of describing and tracing
AI-specific dependencies [
          <xref ref-type="bibr" rid="ref5">5, 23</xref>
          ].
        </p>
        <p>AI software comprises traditional software components alongside AI-specific artifacts such as training
data, model configurations, and inference pipelines. Unlike conventional software, AI systems evolve
continuously (e.g. in response to data drift and concept drift) necessitating dynamic co-versioning
1https://github.com/nqminds/Trusted-AI-BOM/</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Related Work</title>
      <p>registries to ensure transparency and accountability. Unlike static SBOM inventories, AI Bills of
Materials (AIBOMs) must support traceable, evolving records without requiring frequent regeneration.
This section reviews existing AIBOM proposals and discusses how the proposed TAIBOM approach
addresses their limitations in establishing verifiable trust within AI software supply chains.</p>
      <sec id="sec-3-1">
        <title>3.1. Existing AIBOM Proposals</title>
        <p>Several AIBOM solutions have been proposed. Model cards, such as those by Hugging Face [24] and
Google [25], provide metadata on datasets and models but lack comprehensive provenance tracking
(e.g., training data lineage or tampering history). SBOM extensions, including CycloneDX [26] and
SPDX [27], adapt existing SBOM formats for AI, yet their trust guarantees remain limited and often
unverifiable. ML-specific tools such as DVC [ 28] and MLflow [ 29] support lineage tracking for AI
models but do not enforce verifiable trust mechanisms across the full AI development pipeline.</p>
        <p>A more detailed comparative evaluation of these approaches, including their support for provenance,
tamper detection, and dependency analysis, is presented in Section 5.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. TAIBOM: A Trustable AIBOM Solution</title>
        <p>Existing AIBOM solutions lack trust guarantees — there are no established techniques to verify dataset
provenance or ensure model integrity throughout the AI workflow. TAIBOM addresses this gap by
introducing a trust-enabled AIBOM data model that represents key AI artifacts (including Data, Code,
and AI System objects) along with their interconnections and trust relationships. As detailed in Section 4,
TAIBOM integrates cryptographic attestations, integrity verification, and dataset provenance tracking
to establish verifiable trust across AI-enabled software supply chains. By design, TAIBOM adopts
a general and extensible structure to accommodate a wide range of AI-enabled software systems —
including those incorporating frontier AI such as Large Language Models (LLMs) and other forms of
generative AI.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. TAIBOM Architecture</title>
      <p>The main purpose of TAIBOM is to provide a structured and verifiable framework for managing trust
in AI software components, ensuring their provenance, integrity and traceability across the entire AI
software lifecycle.</p>
      <p>Claims and inferences</p>
      <sec id="sec-4-1">
        <title>4.1. TAIBOM Data Model</title>
        <p>The TAIBOM data model consists of multiple interconnected class objects depicted at Figure 1.
Data is a parent class for all datasets encompassing metadata such as name, label, location,
cryptographic hashes, and last access time. TrainingData extends Data, specifically capturing AI training
datasets, while DataPack aggregates multiple TrainingData instances, representing a structured
collection of datasets used in AI model training.</p>
        <p>Code is a parent class for all code-related objects which encapsulates all software artifacts, tracking
their location, cryptographic integrity, and SBOM references. TrainingCode and InferencingCode
extends the parent Code component and are used to identify and describe training and inference
software along with their SBOMs. SBOM structure further integrates with CVE identifiers, ensuring that
known vulnerabilities are explicitly linked to the AI system’s components. Additionally, License
metadata is incorporated to track software licensing requirements.</p>
        <p>AI System is a parent class that encompasses labels, code references, and training datasets. The
TrainedSystem and InferenceSystem inherits from AISystem, serving as the components
responsible for executing model training and inference respectively. The TrainedSystem component
integrates DataPack and TrainingCode, and producing Weights, which represent the learned model
parameters. Weights inherit from Data and are linked to Config, which encapsulates key AI system
parameters, including associated training data and system metadata. The InferenceSystem is the
object of the resulting AI system, composed on Config and InferencingCode, used for designated
inference tasks in the actual deployment.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. TAIBOM Framework</title>
        <p>A fundamental aspect of the described TAIBOM Data Model is ensuring that all critical components,
including training datasets, training code packages, and SBOM descriptors, are cryptographically signed
and versioned. The linkage between SBOM descriptors and corresponding training code establishes a
chain of trust. Once the system undergoes training with a predefined configuration, trained weights,
inference code, and the SBOM descriptor for the SBOM code itself are also signed. As all components
contain signed hashes and traceback from resulting object to the source object, the architecture provides
integrity, traceability, and accountability, mitigating risks associated with unauthorised modification of
data, model weights, code or AI system configurations.</p>
        <p>The TAIBOM operational workflow, illustrated in Figure 2, follows the conventional structure of
machine learning system development and deployment. It is organised into several key phases, with
TAIBOM augmenting each step through additional data collection, attestation, and integrity mechanisms.
The training phase involves the instantiation of a trained AI system that is explicitly tied to versioned
and attested training data and machine learning code, enabling full traceability of all artefacts involved.
The testing, QA, and validation phase assesses the robustness of the trained system, ensuring compliance
with predefined performance and security requirements. Once validated, the deployable AI system
is assembled using tested and attested components to ensure stability and reproducibility. The final
phase involves the deployed trained AI system, which performs real-world inferences while preserving
integrity guarantees and provenance tracking as defined by the TAIBOM attestation and versioning
data model.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. TAIBOM Use Cases</title>
      <p>This section evaluates the TAIBOM framework against most widespread AI documentation and software
composition tools — SPDX [27], CycloneDX [26] and model cards (Google [25], HuggingFace [24]) — by
examining four critical use cases that represent operational challenge in AI lifecycle assurance, security,
and traceability. The analysis contrasts the technical capabilities and design limitations of existing tools
with TAIBOM’s architecture and operational semantics.</p>
      <p>Model cards provide standardised documentation for AI models, including training configurations,
intended use cases, and known limitations. However, they are self-reported and do not enforce integrity
or provenance checks on datasets, training code, or derived models. CycloneDX and SPDX AI extensions
ofer structured software composition formats, enabling dependency visibility and limited vulnerability
referencing. However, these frameworks do not establish cryptographic links between AI components
(e.g., training datasets, trained weights) and their associated metadata, and thus cannot verify whether
components have remained unchanged or trustworthy throughout the development lifecycle.</p>
      <p>In contrast, TAIBOM enforces signed attestations and cryptographic integrity checks across AI
artifacts, including datasets, training code, configuration files, and model weights. This ensures that all
artifacts can be independently validated and traced to their origin, providing a trust dimension that is
currently absent other approaches.</p>
      <p>Below, we detail four representative use cases (UCs) in which TAIBOM’s trust mechanisms ofer
substantial value. Table 1 provides a comparison of TAIBOM, Model Cards, and SPDX/CycloneDX AI
extensions across some of the most common use cases in the AI model lifecycle.</p>
      <sec id="sec-5-1">
        <title>UC 1: Declaring Training Data for Transparency</title>
        <p>AI model training frequently involves datasets compiled from disparate or opaque sources. Without
structured, verifiable records of data composition, organisations face dificulties validating dataset
provenance, understanding licensing implications, or reproducing experimental outcomes.
SPDX. The SPDX standard, originally designed for traditional software, has been extended via SPDX
AI to include certain AI artifacts. Datasets may be represented as generic File or Package elements
using fields such as name, fileName, licenseConcluded, or checksums. However, SPDX AI treats
data as static external files and lacks a domain-specific schema for dataset structure, role (e.g., training
vs. test sets), or provenance tracking.</p>
        <p>CycloneDX. CycloneDX AI allows datasets to be represented as components with type: data,
supporting metadata like externalReferences and hashes. However, the format is flat and does
not support relationships between datasets (e.g., folds or partitions), nor does it express their functional
roles in training workflows.</p>
        <p>Model Cards. Model cards may include narrative descriptions of datasets, their origin, and limitations,
but ofer no structured or verifiable representation of dataset usage.</p>
        <sec id="sec-5-1-1">
          <title>TAIBOM Approach</title>
          <p>TAIBOM defines Data and TrainingData objects, each containing cryptographic hashes, source
URIs, access times, and licensing metadata. Composite datasets are grouped using the DataPack
structure. These objects are cryptographically signed and versioned, and explicitly linked to the training
configuration ( Config) and TrainedSystem components, ensuring verifiable dataset usage.</p>
        </sec>
        <sec id="sec-5-1-2">
          <title>Brief Evaluation Summary</title>
          <p>TAIBOM enables formal, cryptographically verifiable declarations of training datasets, while existing
tools treat datasets as informal metadata or lack semantics for AI-specific data structuring.</p>
        </sec>
      </sec>
      <sec id="sec-5-2">
        <title>UC 2: Assessing Training Data for Poisoning</title>
        <p>Data poisoning introduces malicious or manipulated samples into training datasets, often without
detection. Detecting poisoning requires dataset versioning, reproducibility, and traceability of data
inputs to model outputs.</p>
        <sec id="sec-5-2-1">
          <title>Existing Approaches</title>
          <p>SPDX. While SPDX allows for checksums on files, these are static and not integrated with AI
worklfows. There is no structured support for dataset versioning or lineage tracking.</p>
          <p>CycloneDX. CycloneDX supports hash-based integrity checks and can represent dataset revisions as
separate components. However, there is no systematic way to express temporal lineage or bind datasets
to specific models.</p>
          <p>Model Cards. Model cards lack dataset version tracking or programmatic validation, and are
unsuitable for forensic analysis.
Each TrainingData object includes versioning metadata and a signed cryptographic hash. TAIBOM
records lineage via associations between DataPack, TrainedSystem, and Weights. Reused datasets
can be detected via hash comparisons, and deviations trigger verification failures. This supports
trace-based identification of potential poisoning across training runs.</p>
        </sec>
        <sec id="sec-5-2-2">
          <title>Brief Evaluation Summary</title>
          <p>TAIBOM supports integrity verification and historical comparison of training datasets — capabilities
absent from SPDX AI, CycloneDX, and model card approaches.</p>
        </sec>
      </sec>
      <sec id="sec-5-3">
        <title>UC 3: Detecting Training and Inference System Tampering</title>
        <p>Post-training tampering of AI components, such as code, configurations, or model weights, can result
in erroneous or malicious behaviour. Mitigation requires strong binding between training inputs and
deployed inference artifacts.</p>
        <sec id="sec-5-3-1">
          <title>Existing Approaches</title>
          <p>SPDX. SPDX AI documents software dependencies and licenses, but does not model training or
inference workflows. There is no construct to relate code to resulting models or configurations.
CycloneDX. CycloneDX AI introduces AI-specific components but lacks explicit workflow modeling.
Dependencies may be represented via a graph, but without semantic links between artifacts (e.g.,
weights generated from training code).</p>
          <p>Model Cards. Model cards describe configurations and limitations, but are static and decoupled from
runtime code and outputs.</p>
        </sec>
        <sec id="sec-5-3-2">
          <title>TAIBOM Approach</title>
          <p>TAIBOM introduces structured workflow representations via TrainedSystem and InferenceSystem
classes. Each references TrainingCode, InferencingCode, Weights, and Config, all of which are
independently signed and versioned. These links enable reproducible revalidation of deployed systems
and detection of unauthorized changes to any component.</p>
        </sec>
        <sec id="sec-5-3-3">
          <title>Brief Evaluation Summary</title>
          <p>TAIBOM supports system-level tamper detection via cryptographic linkage and provenance tracking,
whereas other tools lack workflow semantics or enforcement mechanisms.</p>
        </sec>
      </sec>
      <sec id="sec-5-4">
        <title>UC 4: Evaluating CVE Impact on Training and Inference Systems</title>
        <p>New CVEs may afect software libraries used in AI training or inference. Identifying afected models
requires understanding which code components were used and how they relate to model artifacts.</p>
        <sec id="sec-5-4-1">
          <title>Existing Approaches</title>
          <p>SPDX. SPDX AI allows linking to CVEs via external references, but does not relate these to model
outputs or training processes. Vulnerabilities are scoped to individual files or packages only.
CycloneDX. CycloneDX includes a detailed vulnerability schema, enabling component-level CVE
annotations. However, these annotations are not connected to AI lifecycle stages or model dependencies,
limiting impact analysis.</p>
        </sec>
        <sec id="sec-5-4-2">
          <title>Model Cards.</title>
          <p>ences.</p>
          <p>Model cards do not include CVE information or component-level vulnerability
referTAIBOM binds each TrainingCode and InferencingCode object to a software descriptor
containing CVE and CWE references. These components are directly linked to TrainedSystem and
InferenceSystem objects, allowing automated identification of afected models and inference
systems.
TAIBOM enables propagation of vulnerability information across AI workflows, supporting targeted
risk assessment. Other tools remain limited to surface-level vulnerability declarations without workflow
integration.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Discussion and Limitations</title>
      <p>While TAIBOM ofers a structured and verifiable framework for introducing trust into AI-enabled
systems, several open challenges remain for further research.</p>
      <p>Granularity of Component Descriptions. As with traditional SBOMs, the level of granularity in
TAIBOM can be further refined. Determining the appropriate level of detail for representing AI-specific
artifacts, such as training subsets, model checkpoints, or dynamic configurations, remains an ongoing
area of improvement.
“Trusted” does not necessarily mean “secure”. It is important to clarify that “trusted” in the
context of TAIBOM does not imply complete or absolute security. Rather, it reflects adherence to a
defined trust model that includes cryptographic attestations and provenance guarantees under certain
assumptions — such as the trustworthiness of dataset providers, signing authorities, and the integrity
of the signing infrastructure.</p>
      <p>Trust Chain Recall and Recovery. TAIBOM assumes that components in the supply chain are
verifiable. However, if a component is later found to be compromised (e.g. a poisoned dataset or a
tampered training script), procedures must be in place to recall or revoke the corresponding trust
attestations. Future work will explore mechanisms for restructuring or rebuilding the trust chain in
such scenarios, similar to revocation in certificate-based systems.</p>
      <p>Scalability in Signing Large-Scale Data. Cryptographic signing of large-scale datasets, potentially
petabytes in size, presents practical challenges. To address this, probabilistic or representative sampling
methods (e.g., signing hashed subsets of data blocks) can be employed, enabling eficient integrity
verification without the need to sign entire datasets. However, the use of stronger trust anchors
remains an open consideration, particularly for deployments requiring higher assurance. Balancing the
integration of such mechanisms with performance, scalability, and operational constraints is essential
for practical adoption, especially in power-constrained environments.</p>
    </sec>
    <sec id="sec-7">
      <title>7. Conclusions</title>
      <p>This paper introduced TAIBOM, a trust-enabled AI Bill of Materials framework that extends conventional
SBOM principles with AI-specific constructs. TAIBOM ofers a structured data model, cryptographic
attestation mechanisms, and a provenance-aware architecture to support trust, transparency, and
integrity across the AI pipeline — from training data and model weights to deployed inference systems.</p>
      <p>Through detailed use cases and a comparative evaluation against existing solutions such as Model
Cards, SPDX AI, and CycloneDX AI, we demonstrated TAIBOM’s capacity to address emerging
challenges in AI assurance, security, and compliance. By supporting both technical verification and
auditability, TAIBOM enables principled trust-building in the next generation of AI-enabled software
systems. In addition to AI-specific provenance and integrity guarantees, TAIBOM’s design can also
help detect and mitigate broader system-level issues, such as data corruption during transmission or
storage, and reduce the impact of human errors in artifact handling.</p>
      <p>Future work will focus on advancing the scalability of attestation mechanisms, refining trust recovery
strategies, and integrating TAIBOM with industry-standard development pipelines to support broader
adoption.</p>
    </sec>
    <sec id="sec-8">
      <title>Acknowledgements</title>
      <p>TAIBOM is supported by InnovateUK under grant 10092977: TAIBOM – Trusted AI Bill of Materials.</p>
    </sec>
    <sec id="sec-9">
      <title>Declaration on Generative AI</title>
      <p>Generative AI (ChatGPT, OpenAI) was used primarily for sentence polishing and grammar check in
selected parts of the manuscript to improve fluency. Following the use of this tool, the content was
further reviewed and edited as appropriate.
and Social Change 181 (2022) 121763. URL: https://www.sciencedirect.com/science/article/pii/
S0040162522002888. doi:https://doi.org/10.1016/j.techfore.2022.121763.
[9] F. Gille, A. Jobin, M. Ienca, What we talk about when we talk about trust: Theory of trust for
ai in healthcare, Intelligence-Based Medicine 1-2 (2020) 100001. URL: https://www.sciencedirect.
com/science/article/pii/S2666521220300016. doi:https://doi.org/10.1016/j.ibmed.2020.
100001.
[10] N. Telecommunications, I. Administration, Software bill of materials (sbom), 2021. URL: https:
//www.ntia.gov/page/software-bill-materials, accessed: 2024-08-27.
[11] NIST, Common Platform Enumeration (CPE), 2024. URL: https://nvd.nist.gov/products/cpe,
[Accessed: Jun. 19, 2024].
[12] NIST, Nist’s cve process, 2024. URL: https://nvd.nist.gov/general/cve-process, [Accessed: Jun. 19,
2024].
[13] M. Corporation, About common weakness enumeration, 2024. URL: https://cwe.mitre.org/about/
index.html, [Accessed: Jun. 19, 2024].
[14] M. Feng, Z. Yuan, F. Li, G. Ban, Y. Xiao, S. Wang, Q. Tang, H. Su, C. Yu, J. Xu, A. Piao, J. Xue,
W. Huo, B2sfinder: detecting open-source software reuse in cots software, in: Proceedings of the
34th IEEE/ACM International Conference on Automated Software Engineering, ASE ’19, IEEE
Press, 2020, p. 1038–1049. URL: https://doi.org/10.1109/ASE.2019.00100. doi:10.1109/ASE.2019.
00100.
[15] R. Duan, A. Bijlani, M. Xu, T. Kim, W. Lee, Identifying open-source license violation and 1-day
security risk at large scale, in: Proceedings of the 2017 ACM SIGSAC Conference on Computer
and Communications Security, CCS ’17, Association for Computing Machinery, New York, NY,
USA, 2017, p. 2169–2185. URL: https://doi.org/10.1145/3133956.3134048. doi:10.1145/3133956.
3134048.
[16] A. Hemel, K. T. Kalleberg, R. Vermaas, E. Dolstra, Finding software license violations through
binary code clone detection, in: Proceedings of the 8th Working Conference on Mining Software
Repositories, MSR ’11, Association for Computing Machinery, New York, NY, USA, 2011, p. 63–72.</p>
      <p>URL: https://doi.org/10.1145/1985441.1985453. doi:10.1145/1985441.1985453.
[17] Trivy: Open source vulnerability scanner, 2024. URL: https://trivy.dev/v0.33/, accessed: 2024-11-28.
[18] Microsoft, Sbom tool: Generate software bill of materials (sboms), n.d. URL: https://github.com/
microsoft/sbom-tool, accessed: 2024-11-28.
[19] Anchore, Syft: A cli tool and library for generating sboms from container images and filesystems,
n.d. URL: https://github.com/anchore/syft, accessed: 2024-11-28.
[20] W. Tang, Z. Xu, C. Liu, J. Wu, S. Yang, Y. Li, P. Luo, Y. Liu, Towards understanding third-party
library dependency in c/c++ ecosystem (2022). URL: https://doi.org/10.1145/3551349.3560432.
[21] T. Hu, Z. Xu, Y. Fang, Y. Wu, B. Yuan, D. Zou, H. Jin, Fine-grained code clone detection with
blockbased splitting of abstract syntax tree, in: Proceedings of the 32nd ACM SIGSOFT International
Symposium on Software Testing and Analysis, ISSTA 2023, Association for Computing Machinery,
New York, NY, USA, 2023, p. 89–100. URL: https://doi.org/10.1145/3597926.3598040. doi:10.1145/
3597926.3598040.
[22] S. Woo, S. Park, S. Kim, H. Lee, H. Oh, Centris: A precise and scalable approach for identifying
modified open-source software reuse, in: Proceedings of the 43rd International Conference
on Software Engineering, ICSE ’21, IEEE Press, 2021, p. 860–872. URL: https://doi.org/10.1109/
ICSE43902.2021.00083. doi:10.1109/ICSE43902.2021.00083.
[23] Q. Lu, L. Zhu, X. Xu, J. Whittle, Z. Xing, Towards a roadmap on software engineering for responsible
ai, in: Proceedings of the 1st International Conference on AI Engineering: Software Engineering
for AI, CAIN ’22, Association for Computing Machinery, New York, NY, USA, 2022, p. 101–112.</p>
      <p>URL: https://doi.org/10.1145/3522664.3528607. doi:10.1145/3522664.3528607.
[24] H. Face, Model cards, https://huggingface.co/docs/hub/model-cards, 2020. Accessed: 2025-03-31.
[25] M. Mitchell, S. Wu, A. Zaldivar, P. Barnes, L. Vasserman, B. Hutchinson, E. Spitzer, I. D. Raji,
T. Gebru, Model cards for model reporting, in: Proceedings of the Conference on Fairness,
Accountability, and Transparency (FAT* 2019), ACM, 2019, pp. 220–229. doi:10.1145/3287560.
3287596.
[26] CycloneDX, Cyclonedx sbom standard, https://cyclonedx.org, 2024. Accessed: 2025-03-31.
[27] S. Project, Spdx ai: Extending the software package data exchange for ai systems, https://spdx.</p>
      <p>dev/specifications/ai/, 2023. Accessed: 2025-03-31.
[28] DVC, Data version control, https://dvc.org, 2020. Accessed: 2025-03-31.
[29] Accelerating the machine learning lifecycle with MLflow, USENIX Association, Santa Clara, CA,
2019.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>T.</given-names>
            <surname>Stalnaker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Wintersgill</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Chaparro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. Di</given-names>
            <surname>Penta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. M.</given-names>
            <surname>German</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Poshyvanyk</surname>
          </string-name>
          ,
          <article-title>Boms away! inside the minds of stakeholders: A comprehensive study of bills of materials for software systems</article-title>
          ,
          <source>in: Proceedings of the IEEE/ACM 46th International Conference on Software Engineering</source>
          , ICSE '24,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2024</year>
          . URL: https://doi.org/10. 1145/3597503.3623347. doi:
          <volume>10</volume>
          .1145/3597503.3623347.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>W.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Synovic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Sethi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Indarapu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hyatt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. R.</given-names>
            <surname>Schorlemmer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. K.</given-names>
            <surname>Thiruvathukal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. C.</given-names>
            <surname>Davis</surname>
          </string-name>
          ,
          <article-title>An empirical study of artifacts and security risks in the pre-trained model supply chain</article-title>
          ,
          <source>in: Proceedings of the 2022 ACM Workshop on Software Supply Chain Ofensive Research and Ecosystem Defenses</source>
          , SCORED'22,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2022</year>
          , p.
          <fpage>105</fpage>
          -
          <lpage>114</lpage>
          . URL: https://doi.org/10.1145/3560835.3564547. doi:
          <volume>10</volume>
          .1145/3560835.3564547.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>J.</given-names>
            <surname>Biden</surname>
          </string-name>
          ,
          <article-title>Executive order on improving the nation's cybersecurity</article-title>
          , https://www.whitehouse.gov/briefing-room/presidential-actions/
          <year>2021</year>
          /05/12/ executive
          <article-title>-order-on-improving-the-nations-</article-title>
          <string-name>
            <surname>cybersecurity</surname>
            <given-names>/</given-names>
          </string-name>
          ,
          <year>2021</year>
          . Accessed:
          <fpage>2024</fpage>
          -06-20.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>D. C.</given-names>
            <surname>European Commission</surname>
          </string-name>
          , Cyber resilience act,
          <year>2023</year>
          . URL: https://www.cisa.gov/sites/default/ ifles/2023-09/EU%20Commission%
          <article-title>20SBOM%20Work_508c</article-title>
          .pdf, accessed:
          <fpage>2024</fpage>
          -08-27.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>B.</given-names>
            <surname>Xia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Bi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Xing</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Lu</surname>
          </string-name>
          ,
          <string-name>
            <surname>L. Zhu,</surname>
          </string-name>
          <article-title>An empirical study on software bill of materials: Where we stand and the road ahead</article-title>
          ,
          <year>2023</year>
          . URL: https://arxiv.org/abs/2301.05362. arXiv:
          <volume>2301</volume>
          .
          <fpage>05362</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>N.</given-names>
            <surname>Zahan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Tamanna</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Enck</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Williams</surname>
          </string-name>
          ,
          <article-title>Software bills of materials are required. are we there yet?</article-title>
          ,
          <source>IEEE Security &amp; Privacy</source>
          <volume>21</volume>
          (
          <year>2023</year>
          )
          <fpage>82</fpage>
          -
          <lpage>88</lpage>
          . doi:
          <volume>10</volume>
          .1109/MSEC.
          <year>2023</year>
          .
          <volume>3237100</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>H.</given-names>
            <surname>Choung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>David</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. R.</surname>
          </string-name>
          <article-title>and, Trust in ai and its role in the acceptance of ai technologies</article-title>
          ,
          <source>International Journal of Human-Computer Interaction</source>
          <volume>39</volume>
          (
          <year>2023</year>
          )
          <fpage>1727</fpage>
          -
          <lpage>1739</lpage>
          . URL: https://doi.org/10.1080/10447318.
          <year>2022</year>
          .
          <volume>2050543</volume>
          . doi:
          <volume>10</volume>
          .1080/10447318.
          <year>2022</year>
          .
          <volume>2050543</volume>
          . arXiv:https://doi.org/10.1080/10447318.
          <year>2022</year>
          .
          <volume>2050543</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>N.</given-names>
            <surname>Omrani</surname>
          </string-name>
          , G. Rivieccio,
          <string-name>
            <given-names>U.</given-names>
            <surname>Fiore</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Schiavone</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. G.</given-names>
            <surname>Agreda</surname>
          </string-name>
          ,
          <article-title>To trust or not to trust? an assessment of trust in ai-based systems: Concerns, ethics and contexts</article-title>
          , Technological Forecasting
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>