Lakera Guards Inputs Without Governing System Confidence

by Nick Clark | Published March 28, 2026 | PDF

Lakera provides real-time detection of prompt injection attacks, data leakage attempts, and toxic content targeting LLM applications. The platform evaluates each input for adversarial patterns and blocks threats before they reach the model. The threat detection is fast, accurate, and addresses a genuine security need. But defending against individual adversarial inputs does not govern the system's overall operational confidence. A system under sustained attack, where threat detection is blocking an increasing proportion of inputs, should reduce its execution authority rather than continuing to process the inputs that pass through the filter. Confidence governance provides this: persistent state that integrates threat detection patterns into a computation that modulates execution authority based on the threat landscape trajectory.


1. Vendor and Product Reality

Lakera AI, founded in Zurich in 2021 by David Haber, Mateo Rojas-Carulla, and Matthias Kraft and now operating with offices in Zurich and San Francisco, is among the most visible pure-play LLM-security vendors in the post-ChatGPT enterprise market. Its flagship product, Lakera Guard, exposes a low-latency REST and Python-SDK interface that classifies inputs and outputs to large language model applications in real time, returning verdicts on prompt injection, jailbreak attempts, personally identifiable information leakage, training-data extraction, toxic content, and policy-defined unsafe categories. The detection backbone is informed by Gandalf — Lakera's gamified red-team platform that has accumulated tens of millions of adversarial-prompt interactions from public players — which provides labeled corpora for the underlying classifier ensemble.

The product surface fits cleanly into the architectural slot enterprises have come to expect for "AI security": a sidecar evaluator placed in front of, behind, or surrounding the model call. Customers integrate Guard into chatbots, retrieval-augmented generation applications, agentic workflows, and copilots through a few lines of middleware that route the user prompt and the model response through Lakera's hosted endpoint, receive a structured threat verdict, and proceed, block, or transform accordingly. Lakera Red supplies an automated red-teaming counterpart for pre-production assurance, and Lakera Chrome Extension extends per-prompt evaluation to the consumer-tooling layer. The customer base concentrates in regulated verticals adopting LLM applications cautiously — financial services, insurance, healthcare, defense — and in technology firms that ship LLM-powered features at consumer scale.

Within its scope, Lakera is technically rigorous and operationally credible. The latency budget is tight enough for synchronous integration, the policy taxonomy is configurable, the threat-intelligence loop driven by Gandalf is genuinely differentiated, and the company has been a visible contributor to OWASP's LLM Top 10 and to industry conversation about adversarial input taxonomy. The product is, in effect, the reference implementation of the input-and-output guardrail pattern for LLM applications. What it deliberately does not do — by architectural design, not by oversight — is maintain a persistent, system-wide confidence state for the LLM application as a whole.

2. The Architectural Gap

The structural property Lakera does not exhibit is governed execution authority over the LLM application as a temporally persistent agent. Each input crosses Guard once, is classified, and either passes or is blocked; the verdict is per-call and stateless from the application's standpoint. Lakera's own systems may aggregate telemetry for product improvement, but the application receiving the verdict is given no governed handle on its own operational confidence — only a permit-or-deny flag for the immediate request. This is the same architectural shape as a network intrusion-prevention sensor: high-quality per-packet classification, no governance over the host's own execution posture as the threat landscape shifts.

The gap matters because adversarial pressure against LLM applications is statistical, not categorical. No classifier achieves perfect recall, and the public literature on prompt-injection evasion makes clear that adversaries develop novel surface forms faster than any single detector can be retrained. Under sustained attack, the absolute volume of undetected adversarial inputs reaching the model increases even when detection rate holds constant; meanwhile, social-engineering pressure on the model itself — context poisoning, indirect injection through retrieved documents, multi-turn manipulation — escalates in ways that no per-input verdict can summarize. A defensible AI system in such conditions should not merely block more inputs. It should reduce its own execution authority: narrow its output space, escalate to human oversight, defer reversible actions, and refuse irreversible ones, until the threat trajectory subsides.

Lakera cannot patch this from within Guard's architecture because the product is designed as a stateless per-call evaluator with a managed-service operational model. Adding rate-limit features, dashboards, or per-tenant aggregate metrics produces operational telemetry, not governed confidence state; the application still owns the execution decision, and Lakera still owns only the per-input classification. The closest adjacent capability — anomaly detection over Guard's own verdict stream — would yield monitoring alerts to a SOC, not a structurally credentialed confidence variable that the LLM application can read and that modulates its own actuation behavior. The chain from threat signal to execution-authority modulation simply does not exist as an architectural element of the Lakera product.

3. What the AQ Confidence-Governance Primitive Provides

The Adaptive Query confidence-governance primitive specifies that an AI system maintain a persistent, multi-input confidence state and that this state structurally modulate the system's execution authority across all actuators. Confidence is computed from a vector of operational signals — threat-detection verdicts, output-quality estimators, model-uncertainty measures, distribution-shift indicators, environmental telemetry, and authority-credentialed observations from upstream governance — composed under a published weighting scheme. The state is temporally integrated, so a sustained negative signal compounds while transient excursions decay; rate-of-change detection identifies escalating patterns, and a differential-alarm channel triggers on sudden landscape shifts.

The output of confidence governance is not a numerical score consumed by a dashboard. It is a graduated execution mode admitted from a defined mode set: full executing, scoped executing, inquiry, deferral, refusal. Each mode is structurally defined: scoped executing narrows the output distribution and disables irreversible tools; inquiry forces the system to ask clarifying questions before acting; deferral routes the request to human oversight with full lineage; refusal returns a governed non-action accompanied by a credentialed explanation. The mode is a property of the system, not of the request; a single confidence transition simultaneously modulates every actuator the system controls, which is the structural difference between confidence governance and per-call gating.

The primitive composes hierarchically. A subagent's confidence state becomes a credentialed observation in the parent agent's chain; a fleet's aggregate confidence is a credentialed observation in coalition-level governance. The trajectory projection — projecting the confidence state forward under current conditions — informs proactive posture changes rather than reactive blocking. The primitive is technology-neutral with respect to the input classifiers (Lakera, Protect AI, Guardrails AI, custom detectors all admit), and is disclosed under the AQ provisional family as a structural condition for operationally governed AI rather than merely defended AI.

4. Composition Pathway

Lakera composes with AQ as a credentialed observation source feeding into the confidence-governance computation. What stays at Lakera: Gandalf's threat-intelligence flywheel, the classifier ensemble, the policy taxonomy, the red-teaming product, the latency-engineered serving infrastructure, and the entire customer relationship for input-and-output guardrails. Lakera's investment in adversarial-prompt research and in low-latency classification remains its differentiated layer; that investment is unreplicable by AQ and is precisely the kind of specialization the architecture is intended to compose with rather than supplant.

What moves to AQ: the persistent confidence state, the execution-mode transitions, the cross-actuator coordination, and the lineage record that ties confidence transitions to the credentialed observations that drove them. Concretely, Lakera Guard is configured to emit each verdict — including soft signals such as classifier margins and per-category likelihoods — as an authority-credentialed observation signed under a Lakera-tenant credential. The AQ confidence engine admits Guard observations alongside other inputs (model log-probability uncertainty, output-quality estimators, retrieval-source provenance, agentic-tool reversibility flags), composes them under a customer-published weighting scheme, and emits a graduated execution mode that the LLM application's actuator layer obeys. The actuator layer in turn does not call irreversible tools while in deferral mode, narrows generation while in scoped-executing mode, and produces lineage records on every transition.

The integration is straightforward operationally. An existing Lakera customer adds an AQ middleware layer that receives Guard verdicts plus the application's own signals and exposes a single confidence-state interface to the LLM application. The customer authors a confidence policy — what weights, what thresholds, what mode transitions — once, and the policy governs every model call the application makes thereafter. Lakera retains its surface; AQ supplies the governance shell. The new commercial territory unlocked for Lakera is the regulated-AI segment that today rejects pure per-call guardrails as insufficient under EU AI Act high-risk classifications, NIS2 incident-response obligations, and SEC cyber-disclosure expectations that increasingly look for system-wide governance rather than point controls.

5. Commercial and Licensing Implication

The fitting arrangement is a credentialed-observation publishing partnership with embedded substrate licensing on the AQ side. Lakera publishes Guard verdicts as signed observations into customer-owned AQ confidence chains; AQ licenses the confidence-governance primitive to enterprise customers as part of an LLM-governance subscription, priced per-credentialed-input-stream or per-application-under-governance rather than per-seat. Lakera's revenue line remains its own; AQ's revenue line is the governance shell that elevates Lakera's signal into operationally governed authority modulation.

What Lakera gains: a structural answer to the increasingly common procurement question of "how does your guardrail integrate with our system-level AI governance," a defensible position against in-platform guardrails from hyperscalers (Azure AI Content Safety, AWS Bedrock Guardrails, Google Vertex AI safety filters) by elevating the architectural floor from per-call gating to governed-system participation, and a forward-compatible posture for high-risk AI Act deployments where pure per-input controls will not satisfy conformity assessment. What the customer gains: portable, vendor-agnostic confidence governance whose authority taxonomy belongs to the customer rather than to Lakera; cross-vendor composition with other detector vendors and with the customer's own internal signals; and a lineage record that satisfies regulatory and forensic reconstruction requirements that per-call logging cannot satisfy. Honest framing — the AQ primitive does not replace Lakera. It gives Lakera the system-level governance shell that per-input guardrails have always needed and never had.

Nick Clark Invented by Nick Clark Founding Investors:
Anonymous, Devin Wilkie
72 28 14 36 01