Stable Diffusion's Training Has No Provenance Layer
by Nick Clark | Published March 27, 2026
Stability AI shipped Stable Diffusion 1.x and 2.x, the SDXL family, Stable Diffusion 3, and Stable Audio as open-weight generative models trained on internet-scale image-text and audio-text corpora. The releases democratized high-quality generation and seeded an entire ecosystem of fine-tunes, LoRAs, and downstream products. They also produced Andersen v. Stability AI and the parallel Getty Images litigation, in which artists and rights holders allege that Stability's training pipeline ingested copyrighted works without authorization and that the resulting models reproduce those works in derivative form. The structural problem visible through that litigation is not whether Stability filtered its datasets. It is that the training pipeline has no provenance layer — no in-pipeline mechanism that traces which training samples influenced which model capabilities at which depth, no memorization detector that prevents specific training images from being reconstructable in output, and no governance interface that lets a rights holder, regulator, or platform operator answer the structural question of attribution. Training governance with depth-selective gradient routing and provenance tracing supplies that missing layer.
Vendor and Product Reality
Stability AI is the company that made open-weight image generation a mainstream reality. The original Stable Diffusion 1.4 and 1.5 releases in 2022 — latent diffusion models trained on subsets of LAION-5B with CLIP text conditioning — were the first publicly downloadable text-to-image systems competitive with DALL-E 2 and Midjourney. SDXL, released in 2023, scaled the latent diffusion architecture to 3.5 billion parameters with a refiner stage and dual text encoders. Stable Diffusion 3, released through 2024 and 2025, pivoted to a multimodal diffusion transformer (MMDiT) architecture with rectified flow training and improved compositional understanding. Stable Audio extended the family into music and sound effect generation through a latent diffusion architecture conditioned on text prompts. Stable Video Diffusion and the 3D generation lines round out the franchise.
The open-weight strategy is structurally important. Stability publishes model weights under licenses that permit research and, with later restrictions, commercial use; the Hugging Face ecosystem distributes the weights; ComfyUI, Automatic1111, and Forge supply local inference UIs; CivitAI hosts hundreds of thousands of LoRA fine-tunes and checkpoint merges; downstream products from Adobe Firefly's competitor tier through indie creative tooling are built on Stability foundations or Stability-derived weights. The training corpora are documented at a high level — LAION-5B and its filtered subsets for the early models, undisclosed but discussed mixtures for SD3 and Stable Audio — and the training compute is in the high single-digit to low double-digit millions of dollars per major release. The technical execution is real.
The legal and commercial overlay is the relevant context. Andersen v. Stability AI, filed in 2023 in the Northern District of California, has progressed past motion-to-dismiss into discovery on direct copyright infringement and induced infringement theories. Getty Images v. Stability AI proceeds in parallel in the United Kingdom and the United States, with allegations that include reproduction of the Getty watermark in generated outputs. The European Union's AI Act imposes training-data transparency obligations on general-purpose AI providers above defined compute thresholds, with the GPAI Code of Practice operationalizing those obligations. The U.S. Copyright Office's 2025 guidance on generative AI training has tightened the fair-use analysis. The commercial environment is one in which open-weight training without provenance has moved from a tolerated industry practice to an active legal exposure.
Architectural Gap
The structural problem is that latent diffusion training, as practiced, is provenance-free by construction. Gradients flow uniformly across the U-Net or MMDiT backbone for every training sample. The optimizer integrates contributions from billions of samples into a single weight tensor without retaining any record of which sample contributed to which weight, at which depth, with what magnitude. By the time training converges, the model is a fused statistical summary of the entire corpus, and the question "which training image contributed to this generated output" has no structural answer. Influence-function approximations and post-hoc attribution methods exist in the research literature, but they are statistical estimators with substantial error bars, not pipeline-native provenance records.
Two related gaps compound the first. The first is memorization. Latent diffusion models with sufficient capacity and sufficient sample repetition will memorize specific training images, and a sufficiently specific prompt — sometimes including the original caption — will reconstruct them. Carlini and collaborators demonstrated this for Stable Diffusion 1.x in published work; the Getty litigation alleges it for watermarked images; the SDXL and SD3 training pipelines have no in-flight detector that prevents memorization from happening in the first place. The second is depth-uncontrolled style ingestion. A particular artist's style, present across hundreds or thousands of training images, is absorbed into the model at whichever depths the gradient happens to route it. There is no mechanism that confines style information to identifiable layers, no mechanism that lets the training pipeline attenuate one artist's influence without retraining the entire network, and no mechanism that produces an auditable record of which styles were learned at which depth.
The consequence is that Stability — and any open-weight image-model trainer in the same architectural family — cannot answer the questions that the legal, regulatory, and commercial environment is now demanding. A rights holder asking whether their work influenced a model output receives a statistical estimate, not a structural answer. A regulator asking for training-data transparency under the EU AI Act receives a dataset description, not a per-capability provenance trace. A downstream licensee asking whether the model can be cleanly purged of a specific training source receives the answer that no, it cannot, because the influence is fused into weights that have no provenance index.
What the Training Governance Primitive Provides
Training governance, in the sense Adaptive Query has patented, is a pipeline-native layer that imposes depth-selective gradient routing, memorization detection, and provenance tracing on the training process itself. Depth-selective gradient routing means that training samples carry category metadata — style, composition, semantics, identity, watermark — and the optimizer routes the gradient contributions of each category to a designated band of layers. Style information lands in style layers; compositional information lands in generalization layers; identity-bearing information is either routed to a sequestered region or rejected at ingestion. The routing is enforced in the training loop rather than hoped for as an emergent property of the architecture.
Memorization detection runs in-flight during training. A continuity evaluator monitors the divergence between the model's output distribution conditioned on a training caption and the training image itself. When the divergence falls below a memorization threshold, the training loop intervenes — by attenuating the gradient on the offending sample, by introducing a regularization term, or by removing the sample from subsequent epochs. The model trained under this regime carries a structural guarantee that no training image is reconstructable from its caption, which is the property that watermark-reproduction allegations directly target.
Provenance tracing produces an indexed record, written during training, that maps each training sample (or each category cohort, depending on the granularity setting) to the layer ranges and weight updates it influenced. The index is queryable post-training. A rights holder asking whether their work influenced a particular model capability receives a structural answer derived from the provenance index, not a statistical estimate. A trainer wishing to attenuate or remove a specific source's influence — under license-compliance, settlement, or regulatory pressure — has a structural target to operate on rather than a uniform weight tensor. The primitive does not eliminate the legal questions, but it converts them from unanswerable into answerable, which is the precondition for any defensible licensing, opt-out, or compliance regime.
Composition Pathway
Training governance composes with Stability's existing pipeline rather than replacing it. The latent diffusion architecture, the MMDiT transition in SD3, the rectified flow training objective, the dual text encoders, and the VAE remain unchanged. The governance layer attaches at the data-loader, the optimizer, and the checkpointing surfaces. The data loader is extended to attach category metadata to each sample, sourced from the existing dataset-curation pipeline that already classifies for safety and quality. The optimizer is replaced or wrapped with a depth-selective variant that consumes the category metadata and routes gradients accordingly. The checkpointing path is extended to write the provenance index alongside the model weights, on the same cadence as existing checkpoints.
For SDXL- and SD3-class training runs, the engineering integration is incremental rather than green-field. The category metadata can be bootstrapped from existing CLIP embeddings and dataset taxonomies. The depth-selective optimizer is a drop-in replacement for AdamW with additional routing tensors. The memorization detector runs as an auxiliary head on the training cluster and adds modest compute overhead — single-digit percentage on representative configurations. The provenance index is a separate artifact whose storage cost is sublinear in dataset size when category-cohort granularity is selected. For Stable Audio, the same primitive applies with audio-domain category schemas (genre, instrumentation, voice identity) replacing the image-domain schemas.
Downstream, the governance primitive enables compliance interfaces that the existing pipeline cannot offer. An opt-out registry — required under the EU AI Act for text-and-data-mining objections — becomes enforceable because the provenance index identifies which model regions an opted-out source influenced. A license-compliance attestation becomes verifiable because the index supports queries against specific source identifiers. A settlement-driven attenuation of a specific artist's influence becomes a targeted weight-region operation rather than a full retrain. The open-weight release model is preserved; the addition is the governance metadata that ships alongside the weights.
Commercial and Licensing
Adaptive Query offers the training governance primitive under licensing terms that recognize Stability's position as a foundational open-weight trainer and the active legal pressure on the open-weight training segment. The licensing structures contemplate a per-training-run license keyed to compute spend, a paid-up enterprise license covering a release family (SD3, SDXL successors, Stable Audio successors), and a co-development pathway in which Stability and Adaptive Query jointly carry the governance primitive into a release that ships with provenance metadata as a first-class artifact. The reference implementation is available as a PyTorch-native optimizer and data-loader extension targeted at the training stacks Stability already operates.
The patent claims cover the training governance architecture, the depth-selective gradient routing mechanism, the in-flight memorization detector, and the provenance-index data structure. They do not encumber the latent diffusion architecture, the MMDiT transformer, the rectified flow objective, or any of the model architectures themselves, all of which remain Stability's to use. The commercial proposition is that the litigation exposure visible in Andersen, Getty, and the EU AI Act compliance regime is converted from an existential overhang into a structural compliance position — the first open-weight trainer to ship a provenance-bearing release defines the licensing baseline that the rest of the segment will be measured against, and recovers the ability to license training data, accept opt-outs, and respond to discovery in a way the current pipeline architecturally cannot.