Skip to main content
AITF M1.8-Art12 v1.0 Reviewed 2026-04-06 Open Access
M1.8 M1.8
AITF · Foundations

Supply-Chain Security for ML Dependencies and Model Weights

Supply-Chain Security for ML Dependencies and Model Weights — AI Use Case Management — Foundation depth — COMPEL Body of Knowledge.

9 min read Article 12 of 15

This article walks the AI supply chain at production grade, the integrity controls that detect compromise, and the AI Bill of Materials practice that makes the chain auditable for both internal governance and external compliance.

The AI supply chain at production grade

A production AI system at any meaningful scale inherits from upstream sources at six points.

Software libraries. The training and inference code depends on a tree of libraries — the ML framework (PyTorch, TensorFlow, JAX), the training utilities, the data-processing libraries, the serving frameworks, the platform-integration libraries, and the transitive dependencies of all of the above. A typical production ML repository depends on hundreds of libraries directly and thousands transitively. The supply-chain risk is the same as for any other software project, with the additional complication that ML libraries often include native code with elevated privileges (compiled CUDA kernels, custom operators) that the typical dependency-scanning tool sees less well into.

Base models. Trained models are increasingly inherited rather than trained from scratch. A computer-vision system may fine-tune from an ImageNet pre-trained backbone; a Large Language Model (LLM) application may fine-tune or prompt a foundation model from a model hub or a hosted provider. The base model is itself a piece of code (in the operational sense — it executes against inputs and produces outputs) that was authored upstream and that the operator must trust. Hugging Face, the major cloud-platform model gardens, and the proprietary catalogs from frontier-model providers are all upstream sources whose integrity the operator depends on.

Pre-trained adapters and embeddings. Beyond complete base models, the operator may inherit smaller artefacts — Low-Rank Adapter weights, embedding tables, tokenizers, fine-tuning datasets distributed alongside models. Each is a vector by which compromise propagates and each typically receives less scrutiny than the headline model artefact.

Training datasets. Datasets purchased from data brokers, licensed from partners, scraped from public sources, or contributed by the open community enter the training pipeline as inputs whose integrity the operator must verify. Article 5 of this module addresses the data poisoning attacks the dataset supply chain enables; this article addresses the artefact-level integrity controls the operator needs.

Container images and infrastructure tooling. The runtime environment in which training and inference execute — base container images, sidecar processes, infrastructure-as-code modules, the cloud-platform managed services themselves — is an upstream-sourced layer whose compromise is invisible to the application-level controls. The Solorigate / SolarWinds incident in 2020 demonstrated the impact of this vector across the broader software industry; the lesson applies to AI infrastructure with the same force.

Cloud-platform managed services. When the operator uses a hosted training service, a hosted model registry, a hosted inference service, or a hosted model API, the operator inherits the platform’s supply-chain posture. The inheritance is not infinitely transferable — the operator remains accountable for the application-level outcome — but it shifts the controls the operator must implement directly versus the controls the operator demands evidence of from the platform.

The MITRE ATLAS knowledge base https://atlas.mitre.org/ catalogs supply-chain compromise as an Initial Access tactic against AI systems and documents real-world cases against the public model-hub ecosystem. The NIST AI Risk Management Framework Cybersecurity profile https://www.nist.gov/itl/ai-risk-management-framework prescribes supply-chain integrity as a managed practice and explicitly contemplates the AI-specific entry points above.

Integrity controls that work

The integrity controls that work for the AI supply chain layer the techniques the broader software supply chain has converged on, with AI-specific extensions where the artefact class warrants.

Software Bill of Materials. Every production ML codebase emits a Software Bill of Materials in a standard format (SPDX, CycloneDX) that enumerates every direct and transitive dependency with version pinning. The SBOM is generated by the build pipeline, signed, and stored alongside the build artefact. SBOM generation is now table-stakes for production software and the AI codebase is no exception.

Dependency scanning. The SBOM feeds continuous dependency scanning that detects newly published vulnerabilities in the libraries the codebase depends on. The scanning tooling (Snyk, Dependabot, the cloud-platform native scanners, the OWASP Dependency-Check tool) integrates into the build pipeline and into the production-monitoring stack so vulnerabilities are detected both at build time and continuously thereafter.

Signed artefacts. Every build artefact — the container image, the model file, the dataset bundle — is signed by the build pipeline using a key the build pipeline holds. Sigstore, in-toto attestations, and the cloud-platform-native signing services all support the pattern. The signature is verified at deployment time, ensuring that the artefact loaded into production is the artefact the build pipeline produced and approved.

Attested builds. Beyond signing the artefact, the build itself is attested: the pipeline records the source commit, the build environment, the dependency tree, and the test results, and signs the attestation. The attestation supports investigation when an upstream compromise is later discovered, allowing the operator to determine which artefacts were built before the compromise (safe), during it (suspect), and after the patch (safe again).

AI Bill of Materials. The Software Bill of Materials does not naturally cover AI-specific artefacts — base models, datasets, adapters, embeddings, tokenizers. The AI Bill of Materials (AI-BOM) extends the SBOM concept to cover them. The AI-BOM enumerates every AI artefact the production system inherits, with the same integrity properties (signed, attested, version-pinned) the SBOM provides for software. AI-BOM is an emerging standard in 2026; the major cloud platforms and the leading governance-tooling vendors are converging on a shared schema and the regulatory community (notably the European Union under the AI Act, Article 15 https://artificialintelligenceact.eu/article/15/) is starting to require AI-BOM evidence for high-risk systems.

Model integrity verification at load time. The deployed inference service verifies the model file’s signature before deserializing it, refusing to load any model that does not pass verification. The control catches both supply-chain compromise (the upstream artefact was tampered with) and downstream tampering (the artefact was modified between build and deployment). The pattern dovetails with the model IP protection discipline from Article 4.

Vendored mirrors for upstream sources. High-stakes operators do not depend directly on public package repositories or model hubs; they mirror the upstream artefacts into a controlled internal repository where they are integrity-verified, scanned, and pinned. The mirror pattern provides a control point at which compromise is detectable and a recovery point if upstream changes break the operator’s pipelines.

NIST SP 800-218A https://csrc.nist.gov/pubs/sp/800/218/a/final prescribes the controls above as Secure Software Development Framework practices for generative AI systems. ISO/IEC 42001:2023 Annex A.7 https://www.iso.org/standard/81230.html requires AI Management System operators to manage the AI supply chain with controls equivalent to those of the broader software supply chain. The Gartner AI TRiSM Hype Cycle https://www.gartner.com/en/articles/gartner-top-strategic-technology-trends-for-2024 tracks the maturity of AI-BOM tooling and the integration with broader software-supply-chain platforms.

Maturity Indicators

Foundational. The organization has no SBOM for its ML codebases, no AI-BOM for its AI artefacts, and no signature verification at deployment time. Models are pulled directly from public hubs without integrity checks. Container images are inherited from public sources without scanning. The team cannot enumerate which upstream sources its production systems depend on.

Applied. SBOM is generated for production ML codebases. Dependency scanning is configured. The team has enumerated the AI artefacts each production system depends on, even if the enumeration is not yet automated. Container images are scanned. Signed artefacts are produced for at least the highest-risk releases.

Advanced. AI-BOM is generated and maintained for every production system. Signed artefacts and attested builds are universal. Model integrity verification at load time is enforced. Vendored mirrors are used for the highest-risk upstream sources. The threat model from Article 1 names supply-chain compromise as a vector and the controls map back to it.

Strategic. AI-BOM is consumed by the risk register, by the audit function, and by the regulatory submission process. Supply-chain monitoring detects newly disclosed vulnerabilities in upstream AI artefacts and triggers response on a tracked schedule. Vendor management for AI suppliers includes contractual SBOM and AI-BOM requirements with audit rights. The supply-chain posture is itself audited on a regular schedule by external specialists. Red-team exercises (Article 11) include supply-chain attack scenarios.

Practical Application

A team that has not addressed the AI supply chain should make three changes this quarter. First, generate an SBOM for the production ML codebases and feed it into a dependency scanner; the scanner will surface known vulnerabilities the team is currently exposed to and produce the prioritized backlog for patching. Second, build the AI-BOM by enumerating every base model, adapter, dataset, and pre-trained artefact each production system depends on, with the source URL and version pinned for each. The exercise alone will surface artefacts the team had forgotten about and sources that have changed under the team’s feet. Third, enable signature verification at the inference loader so the model file loaded into production is verifiably the file the build pipeline produced; the change is a small code modification with substantial defensive value.

These three actions create the supply-chain artefacts on which dependency monitoring, vendored mirrors, and the broader maturity progression are subsequently built. They also produce the audit evidence that compliance frameworks (Article 15) increasingly require for AI workloads.


© FlowRidge.io — COMPEL AI Transformation Methodology. All rights reserved.