LangChain Built the Agent Framework. It Did Not Define What an Agent Is.
by Nick Clark | Published March 27, 2026
LangChain became the dominant framework for building LLM-powered agents by providing chains, tools, memory abstractions, and retrieval integrations. The ecosystem is vast and the developer mindshare is enormous. But LangChain agents have no canonical schema. There is no structural definition of what an agent is, what fields it must carry, or how governance, memory, lineage, and identity relate to each other. Agents are assembled from components — they are not structurally defined. The gap is between agent tooling and agent definition. This article positions LangChain against the AQ agent-schema primitive disclosed under provisional 64/049,409.
1. Vendor and Product Reality
LangChain, founded in 2022 by Harrison Chase as an open-source Python project and now operating as LangChain Inc. with a commercial platform (LangSmith for observability and evaluation, LangGraph for stateful multi-agent orchestration, LangServe for deployment) on top of the open-source core, became, within twelve months of launch, the dominant developer framework for LLM-powered application construction. The numbers are real: tens of thousands of GitHub stars, hundreds of thousands of weekly PyPI downloads across the langchain, langchain-core, langchain-community, and langchain-openai packages, integrations with effectively every commercially available foundation model and embedding provider, and a tutorial-and-content ecosystem that has trained a generation of LLM developers.
The architectural shape is well-understood. LangChain provides the chain abstraction (a sequence of LLM calls and transformations), the tool-use pattern (function-calling against a registered toolset), retrieval-augmented generation through a vector-store and document-loader ecosystem, prompt templating, output parsing, and a memory-module library (conversation buffer, summary memory, entity memory, vector-store-backed memory, and others). Agents are assembled by composing these pieces: an LLM, a registered toolset, a prompt template defining the agent's behavioral envelope, optionally a memory module, and a runtime loop (ReAct, OpenAI functions, structured chat, or LangGraph's state-machine model) that decides which tool to call next based on the LLM's reasoning over the current state.
The contribution to making LLM agents accessible is substantial and broadly acknowledged. The chain abstraction, tool-use patterns, and integration ecosystem accelerated the field by a measurable margin. LangSmith provides production-grade observability over agent traces, prompts, and evaluation runs, and LangGraph extends the model into multi-agent orchestration with explicit state graphs. The customer base spans Fortune 500 enterprises adopting LangSmith for internal generative-AI deployments, AI-native startups building vertical agent products, and the long tail of internal R&D teams prototyping agentic workflows. Within its scope as a developer framework and observability platform, the product is rigorous and well-engineered.
2. The Architectural Gap
The structural property LangChain does not provide is a canonical agent schema. A LangChain agent is composed from an LLM reference, a tool list, a prompt template, and optionally a memory module. The composition pattern is flexible by design — different applications assemble different components — but the consequence is that there is no schema that says "an agent must have these fields, they must have these types, and they must relate to each other in these ways." Different agents in the same application can have entirely different structures: one with memory, another without; one with persistent identity across sessions, another spawned ephemeral; one with a defined governance policy reference, another with no policy at all; one with audit-grade lineage of every tool call and memory write, another with no lineage beyond the LangSmith trace.
The gap matters because the absence of a canonical schema makes agents non-interoperable in a structural sense. An agent built by one team cannot be validated by another team's platform because there is no shared definition of what is being validated. Memory is an optional add-on rather than a required typed field, so a platform cannot assume memory exists, cannot assume it has a particular structure, and cannot enforce constraints on how it evolves. Identity is an emergent property of session handling rather than a typed field, so two interactions with what the user thinks of as "the same agent" can carry no structural guarantee that the same identity is at stake. Governance is a prompt-engineering convention — "you are an agent that does X and never does Y" — embedded in natural-language system prompts rather than a credentialed policy reference that the runtime can enforce. Lineage is observability, which is to say a side effect of the LangSmith integration, not a structural property of the agent itself.
The consequence at the industry level is that agents built with LangChain (and with its peers — LlamaIndex, Haystack, AutoGen, CrewAI, the Anthropic Computer Use SDK, and the OpenAI Assistants API) are ad-hoc compositions whose conformance to any safety, governance, or interoperability requirement is an artifact of the team that built them rather than an architectural property. Regulators asking "what governance does this agent operate under, what memory does it carry, what identity is it bound to, and what lineage does it emit" get a tutorial answer ("we use these LangChain modules") rather than a schema-validated answer. LangChain cannot patch this from within its current architecture because the framework's value proposition is exactly the flexibility of composition; imposing a canonical schema across the ecosystem is not a tuning parameter but a substrate property that has to live below the framework, not within it.
3. What the AQ Agent-Schema Primitive Provides
The Adaptive Query agent-schema primitive specifies that every agent in a conforming system carry a canonical, typed-field schema. The required fields are: identity (a credentialed, persistent reference within a published authority taxonomy, distinct from session and from runtime instance), memory (a typed, lineage-bearing structure with defined entry types — episodic, semantic, procedural, working — each entry carrying its own credentialed provenance), governance (a credentialed reference to the policy under which the agent operates, including admissible mode set, tool authorization, and escalation pathway), capabilities (a typed enumeration of the agent's authorized tools and actions, each with credentialed authorization scope), execution state (a typed snapshot of the agent's current run state, distinguishable from memory and from identity), and lineage (a structurally complete record of every observation admitted, decision made, tool invoked, memory written, and actuation committed, with credentials at every step).
These fields are not optional. They are the structural definition of what an agent is. The schema is technology-neutral with respect to the underlying LLM, the embedding model, the vector store, the tool runtime, and the prompt strategy — any of these can be swapped without changing the schema-conformance of the agent. The schema composes hierarchically: an agent can be a sub-agent of a parent agent, with the parent's governance and identity establishing the credentialed envelope within which the sub-agent operates and the sub-agent's lineage rolling up into the parent's. With a canonical schema in place, agents become structurally interoperable: an agent built by one team can be validated by another team's platform because both agree on what an agent structurally is. Governance becomes enforceable because the platform knows where to find the credentialed policy reference and the runtime cannot dispatch a tool call outside the authorized capability set. Memory becomes auditable because its structure is defined, its entries carry provenance, and a platform can verify integrity, enforce retention, and reconstruct state at any past time. Identity becomes persistent because it is a typed field bound to an authority, not an emergent property of session state. Lineage becomes a structural property of the agent rather than an observability side effect. The inventive step disclosed under USPTO provisional 64/049,409 is the closed identity-memory-governance-capability-state-lineage schema as a structural condition for governance-credentialed semantic agents.
4. Composition Pathway
LangChain integrates with AQ as a domain-specialized agent-construction toolkit and observability surface running over the agent-schema substrate. What stays at LangChain: the chain abstraction, the tool-use patterns, the prompt-template system, the document-loader and vector-store ecosystem, the model-provider integrations, LangSmith's observability and evaluation tooling, and the entire developer-experience commercial relationship. LangChain's investment in framework-specific knowledge — the recipe library, the integration matrix, the developer documentation, the LangSmith trace UX — remains its differentiated layer, and the framework's value proposition (flexibility of composition for application developers) is not diminished by the schema substrate.
What moves to AQ as substrate: the agent itself. A LangChain-constructed agent is instantiated as a schema-conformant agent object whose identity, memory, governance, capabilities, execution state, and lineage are typed fields rather than ad-hoc components. The integration points are well-defined. The LangChain agent constructor emits a schema-conformant agent record at instantiation; tool registrations populate the credentialed capability field; memory modules write into the typed memory field with per-entry provenance; the runtime loop dispatches tool calls only within the credentialed capability set, with each dispatch recorded as a lineage event; LangSmith traces become the developer-facing rendering of the structurally complete lineage rather than a separate observability stream. LangGraph's state-machine model becomes the natural runtime for the execution-state field, with state transitions recorded as credentialed observations in the lineage.
The new commercial surface is regulator-grade and enterprise-grade agent deployment for LangChain customers in regulated industries — financial services, healthcare, government, defense — that need to answer governance, audit, and identity questions about deployed agents structurally rather than procedurally. The schema belongs to the customer's authority taxonomy and the regulator's policy authority, not to LangChain's database, so the agent's audit-grade history is portable across framework migrations, model vendor changes, and platform consolidations. This paradoxically makes LangChain stickier in the enterprise tier, because the framework's developer-experience value is what differentiates its access to a substrate the enterprise actually requires.
5. Commercial and Licensing Implication
The fitting arrangement is an embedded substrate license: LangChain Inc. embeds the AQ agent-schema primitive into the open-source core and into LangSmith and LangGraph, and sub-licenses schema participation to its enterprise customers as part of the platform subscription. Pricing is per-credentialed-agent-identity or per-governed-deployment rather than per-developer-seat or per-trace, which aligns with how regulated and large enterprise customers actually consume agentic AI. The open-source core remains unchanged in its developer-facing API surface; the schema is a substrate property that is invisible to developers who do not need it and structurally load-bearing for enterprises that do.
What LangChain gains: a structural answer to the "how do you govern an agent" question that current prompt-engineering and observability tooling only addresses procedurally, a defensible position against competing frameworks (LlamaIndex, Haystack, AutoGen, CrewAI, the Microsoft Semantic Kernel, the OpenAI Assistants API and its successors) by elevating the architectural floor rather than competing on integration count, and a forward-compatible posture against EU AI Act high-risk system requirements, NIST AI RMF, the SEC AI-disclosure regime, and the converging international consensus on credentialed-lineage requirements for production AI systems. What the customer gains: structural agent governance that survives framework migrations and model-vendor changes, a portable audit-grade history of every agent deployment, a shared definition of "agent" that lets internal platform teams validate third-party-built agents without re-reading their construction code, and a single schema spanning experimentation, staging, and production deployments. Honest framing — the AQ primitive does not replace LangChain; it gives the agent ecosystem the schema it has always needed and never had, and the current ad-hoc composition state of production agents is the predictable cost of its absence.