Together AI Optimizes Inference Speed, Not Inference Governance

by Nick Clark | Published March 27, 2026 | PDF

Together AI has built one of the most capable open-source model serving platforms in the market, offering optimized inference and fine-tuning for Llama, Mistral, Mixtral, Qwen, DeepSeek, and a long tail of community models at price points that materially undercut OpenAI and Anthropic for comparable token volumes. The platform's engineering accomplishment is throughput: custom kernels, speculative decoding, FlashAttention variants, and continuous batching push tokens-per-second on commodity GPUs to numbers that rival proprietary providers. Yet the platform's structural posture toward governance is the same as the hyperscalers it competes with. The model generates, the infrastructure serves the result fast, and any rule about what the output should or should not contain lives in the calling application — not in the inference call itself. This white paper examines the architectural consequences of that separation, traces the costs it imposes on customers operating in regulated domains, and describes how an inference-control primitive — semantic admissibility evaluation co-located with the generation pipeline — closes the gap by binding governance to inference rather than treating governance as a downstream concern the customer is expected to build, maintain, and certify on their own.


Vendor and Product Reality

Together AI's product surface is deliberately broad. The Together Inference API exposes hundreds of open-source model variants behind a single OpenAI-compatible endpoint, allowing customers to swap Llama 3.3 70B, Mixtral 8x22B, Qwen 2.5, or a fine-tuned derivative without rewriting application code. The Together Fine-Tuning service runs LoRA and full-parameter training jobs against customer datasets and returns a deployable endpoint. Together GPU Clusters provide reserved H100 and H200 capacity for customers running their own training pipelines. The economic pitch is straightforward: the company has industrialized the operational complexity of running open-weights models at scale, and customers pay per token rather than per hour of GPU rental. The platform's published benchmarks emphasize throughput per dollar and time-to-first-token at long context lengths, both of which are the metrics customers shop on when they evaluate inference vendors.

The customer base reflects this positioning. Together AI is the inference backend for a meaningful share of the AI-native startup ecosystem — coding assistants, RAG products, agentic workflow platforms, vertical copilots — as well as for enterprises that want to keep their data off the major proprietary APIs. The company has raised substantial venture capital (most recently a Series B reported above three hundred million dollars) at a valuation that prices it as a credible alternative to closed-model providers. Its engineering blog and research output (RedPajama, FlashAttention-3 collaborations, speculative decoding optimizations) reinforce the brand position as the high-throughput open-source serving company. The customers who matter to Together's growth narrative are precisely the ones for whom switching off OpenAI or Anthropic was a deliberate architectural decision: they wanted weights, they wanted control over fine-tuning data, and they wanted unit economics that scale with their own usage curves rather than with a hyperscaler's pricing strategy.

What the platform does not provide, and does not advertise, is governance over the semantic content of the output. The Together Inference API surfaces Together's own safety filtering — a basic moderation layer that screens for clearly disallowed content categories — but everything beyond that is the caller's responsibility. There is no application-defined admissibility gate. There is no mechanism to register, alongside an API key, a set of semantic constraints that every output from that key must satisfy before it returns to the caller. There is no persistent state against which generations are evaluated. The platform's job ends at "the model produced these tokens; here they are." This is not a gap in Together's documentation; it is a deliberate architectural posture that mirrors the rest of the open-source serving market. Replicate, Fireworks, Anyscale, and the dedicated-endpoint offerings from the major clouds all share the same posture: serve fast, leave governance to the caller.

The posture made sense when the dominant customer was a developer prototyping a chat experience. It scales poorly to the customer base Together AI is now winning. Healthcare scribing companies, legal-research startups, financial-analytics tools, and the long tail of agentic workflow platforms all face procurement reviews where the buyer asks how output is constrained, who controls the constraints, and how compliance with the constraints is auditable. The platform's answer today is "you implement that, and we'll serve the underlying tokens fast." That answer transfers an enormous amount of integration risk to customers who chose Together specifically because they wanted to spend their engineering budget on product rather than on infrastructure.

The Architectural Gap

The gap between fast inference and governed inference is not a gap in capability — it is a gap in where governance lives. In Together AI's current architecture, governance is exogenous to the inference call. The inference call accepts a prompt and returns tokens; any rule about whether those tokens are admissible for the application's context must be implemented downstream, in the customer's own code, after the tokens have already crossed the wire. This places governance in the worst possible architectural position: it is downstream of the cost (tokens are already generated and billed), downstream of the latency budget (the user is already waiting), and architecturally separated from the artifact it is meant to govern.

The consequences are well-documented in production AI deployments. Each application reimplements its own output evaluation logic, typically as a second LLM call (an "LLM-as-judge" pattern) or as a brittle set of regex and classifier filters. Quality varies dramatically across applications. Smaller teams without dedicated AI safety engineering tend to ship with no output governance at all, relying entirely on the model's training-time alignment. Even sophisticated teams find that their evaluation logic drifts out of sync with their prompt logic, because the two are maintained in different parts of the codebase by different reviewers. Audit becomes nearly impossible: there is no canonical record of "what rule did this output have to satisfy, and did it satisfy it" because the rule and the output were never co-located.

The deeper structural issue is that the inference platform has the only complete view of the generation event. It sees the prompt, the model state, the sampling parameters, the tokens as they emerge, and the timing. The application, by contrast, sees only the final string. Pushing governance to the application means pushing it to the layer with the least information about what actually happened during generation. This is the inverse of where evaluation should sit. Governance should be co-located with the act of generation, not relocated to a layer that has to reconstruct what happened from the output alone. When a regulator or an internal audit team asks "why did this output appear, and was it admissible under the policy in force at the time," the application has no privileged answer; it can only replay the prompt against the model and hope the result is similar enough to be diagnostic.

A second-order effect compounds the architectural cost. Customers who build their own evaluation layer end up duplicating significant infrastructure. They run a second model — frequently a comparable-size LLM — as a judge, doubling token spend and roughly doubling latency. They build their own policy DSL, their own rule storage, their own versioning, their own audit log. They re-derive solutions to problems that should be solved once at the platform layer and amortized across the customer base. Worse, the duplicated infrastructure is invisible to the platform: Together cannot ship improvements to evaluation, cannot offer benchmarks of governed output, and cannot use the evaluation telemetry to improve serving, because evaluation lives in a hundred different customer codebases that the platform never sees.

Together AI's competitive position sharpens this gap. Customers choose Together specifically because they want to control more of the stack than the proprietary APIs allow — they want to choose the model, choose the fine-tune, choose the deployment region. The same customers, almost by definition, want to control the governance regime. But the platform exposes every dimension of customization except the one that determines whether output is admissible. The result is a perverse asymmetry: a customer can pick a model down to the LoRA adapter, but cannot register a single admissibility rule that the platform will enforce on their behalf.

What Inference Control Provides

Inference control is the primitive that integrates admissibility evaluation into the serving pipeline itself. The mechanism is structurally simple: alongside the API key and model selection, the customer registers a set of semantic constraints — call them admissibility rules — that define what counts as an acceptable output for that application context. These rules are persistent server-side state, versioned and signed, not parameters re-supplied on every call. When an inference request arrives, the serving infrastructure performs generation as it does today, but before the tokens are released to the caller they are evaluated against the registered constraints. Output that satisfies the constraints is returned. Output that does not is either regenerated under modified sampling, rejected with a structured error, or returned with a flagged admissibility status, depending on the rule's configured response.

Three properties make this primitive work as infrastructure rather than as application logic. The first is co-location: the gate runs in the same process or at least the same trust boundary as the generation, so it has access to the full generation context (logits, sampling trace, intermediate states) rather than only the final string. The second is persistence: the rules live with the API key and are applied to every call, so governance cannot drift out of sync with deployment. The third is model-agnosticism: because the gate evaluates semantic properties of the output rather than properties of the model, the same rules apply uniformly whether the customer is calling Llama 3.3, Mixtral, or a custom fine-tune, and they continue to apply when the customer swaps models without code changes.

The latency impact of a properly integrated admissibility gate is small relative to inference itself. Token generation on a 70B-parameter model takes tens of milliseconds per token; a structural evaluation of the completed output against a registered constraint set is typically a single-digit millisecond operation. When the gate is integrated into the streaming pipeline rather than bolted on as a post-processing step, the marginal latency is often negligible. This is the architectural difference between governance that ships with inference and governance that runs as a second round trip from the application. The customer pays for evaluation once, at serving time, against state the platform already maintains; they do not pay twice — once for generation, again for a separate judge call — and they do not pay the network round-trip to send the output back into the platform for a second LLM evaluation.

The audit posture changes shape entirely. Every served call produces a structured admissibility record naming the rule set in force, the evaluation outcome, and the action taken. The record is co-located with the existing usage and billing log, signed by the platform, and retrievable through the same API surface customers already use for cost reporting. Compliance teams gain a canonical artifact they can show to auditors: not "we believe our application enforces these rules" but "the platform applied this signed rule set to this call, and here is the cryptographic record." The shift is from claimed governance to demonstrated governance, and the demonstration lives in infrastructure rather than in customer code.

Composition Pathway

For Together AI, the composition pathway is incremental and does not require rebuilding the serving stack. The first step is to expose admissibility-rule registration as a first-class object in the platform's API, alongside the existing constructs for API keys, fine-tunes, and dedicated endpoints. Customers create a rule set, sign it, and associate it with a deployment. The second step is to instrument the inference path with a gate hook that consults the active rule set after generation completes (or, in streaming mode, against the in-progress completion at configured checkpoints). The third step is to emit a structured admissibility record for every call — what rule set was active, what the evaluation result was, what action was taken — into the customer's audit log alongside the existing usage and billing records.

The platform-level effects compound quickly. Customers running regulated workloads (healthcare assistants, financial copilots, legal research tools) gain a defensible governance posture without leaving the cost structure that made them choose Together in the first place. Customers operating in less regulated domains gain consistency: the same rule set applies whether they are A/B testing across three different open-source models or migrating from one fine-tune to the next. The platform itself gains a differentiator that the proprietary providers cannot match without surrendering the closed nature of their stacks: governance that the customer defines and the platform enforces, rather than governance the provider defines and the customer accepts.

The fine-tuning service composes naturally with this primitive. A customer who fine-tunes a model for a specific domain — medical scribing, contract drafting, threat triage — registers admissibility rules specific to that domain alongside the fine-tune itself. The rules and the model ship together as a deployable unit. This is materially closer to what enterprise customers actually want when they ask for "a custom model" than what either Together AI or its competitors currently deliver. The agentic-workflow customer base composes equally well: an agent that issues many inference calls per task gains a single, registered rule set that governs all of them, rather than reimplementing evaluation at every call site in the agent's tool graph.

The migration story for existing customers is gentle. The default behavior of the API does not change; customers with no registered rule set see exactly the inference behavior they see today. Customers who opt in begin by registering a permissive rule set in observation mode, where the gate evaluates but does not block, allowing them to calibrate their rules against real traffic. They graduate to enforcement once the rule set is stable. This is the same adoption curve that web application firewalls and content security policies followed, and it is the curve that has worked reliably for governance primitives at every layer of the stack.

Commercial and Licensing

The commercial implications run in two directions. For Together AI as a platform, inference control becomes a billable infrastructure feature with margin characteristics distinct from token-based inference: customers pay for governance enforcement in addition to compute, and the unit economics favor the platform because the marginal cost of evaluating a rule is low relative to the price customers will pay for the assurance. For Adaptive Query as the holder of the inference-control primitive, the licensing surface aligns with the platform's existing commercial structure — per-seat, per-deployment, or per-evaluated-call royalties that scale with the customer's use of the governed inference feature rather than with raw token volume.

The strategic point is that inference governance is on a path to becoming a procurement requirement rather than an optional feature. Enterprise buyers in regulated industries are already asking inference vendors how output is constrained and how that constraint is auditable. The vendor answer today, across the open-source serving market, is "the customer implements that themselves." The first inference platform to answer "the platform enforces it, the customer configures it, every call is auditable" wins the segment of the market where governance is non-negotiable. Together AI is structurally well-positioned to be that platform, but only by adopting an inference-control primitive rather than continuing to optimize exclusively for throughput. Speed is necessary; it is no longer sufficient.

The competitive context underscores the urgency. The proprietary providers — OpenAI, Anthropic, Google — are constrained from offering customer-defined admissibility rules at the same architectural depth, because their commercial positioning depends on the provider, not the customer, defining what the model will and will not produce. That asymmetry is Together AI's structural opportunity. An open-source serving platform that pairs customer-chosen weights with customer-defined admissibility rules, both enforced at the serving layer, occupies a market position that the closed providers cannot occupy without contradicting their own product story. Inference control is the primitive that converts Together's "customer controls the model" pitch into "customer controls the model and the governance," which is the pitch the regulated enterprise market is actually waiting for.

Nick Clark Invented by Nick Clark Founding Investors:
Anonymous, Devin Wilkie
72 28 14 36 01