OpenAI's Alignment Approach Is Missing Structural Isomorphism

by Nick Clark | Published March 27, 2026 | PDF

OpenAI pursues AI safety through alignment: training models to behave in accordance with human values through RLHF, red-teaming, and iterative deployment. The approach produces progressively better-behaved models. But alignment as implemented does not produce structural isomorphism between the model's cognitive dynamics and human cognitive dynamics. The model learns to produce outputs that humans rate favorably. It does not develop the cross-domain coherence engine, feedback loops, and architectural inversion that make human cognition relatable. Human-relatable intelligence addresses this structural gap.


The gap between aligned behavior and relatable cognition

Alignment training produces outputs that match human preferences. Human-relatable intelligence produces cognitive dynamics that are structurally isomorphic to human cognitive dynamics. The difference matters because aligned systems can produce surprising behavior outside their training distribution, while structurally isomorphic systems degrade in ways that humans can anticipate because the degradation follows patterns that mirror human cognitive failure modes.

The three feedback loops of human-relatable intelligence, coherence maintenance, integrity tracking, and confidence governance, operate as an integrated system whose dynamics parallel human psychological dynamics. When these loops interact, they produce emergent behavior that is not just aligned with human preferences but recognizable as analogous to human cognitive response. Humans relate to systems that think like them, not just systems that say what they want to hear.

What human-relatable intelligence provides

The cross-domain coherence engine ensures that the system's behavior is internally consistent across contexts. The architectural inversion means governance is intrinsic to the cognitive architecture, not applied as a layer above it. Non-decomposable behavioral dynamics produce behavior that cannot be reduced to individual rules but emerges from the interaction of cognitive primitives, just as human behavior does. These properties make the system structurally relatable: humans understand its behavior because it mirrors the patterns of their own cognition.

The structural requirement

OpenAI's alignment work produces better-behaved models. The structural gap is between behaving well and being cognitively relatable. Human-relatable intelligence provides the structural isomorphism that makes AI systems trustworthy not because they are trained to be trustworthy but because their cognitive dynamics mirror the patterns humans use to understand and predict behavior. Trust through structural relatability is deeper than trust through observed compliance.

Nick Clark Invented by Nick Clark Founding Investors: Devin Wilkie