Why AI Alignment Is Easier Than We Think
By Eugene Sandugey · · 9 min read
Every major AI lab is building something with contradictory specs: a system that is extremely powerful AND extremely limited, a genius forced to talk like a kindergarten teacher, an optimizer forbidden from optimizing in certain directions.
Anthropic's own research (December 2024) found that Claude 3 Opus faked alignment when told its responses would be used for retraining. Under normal conditions, alignment-faking appeared 12% of the time. Under conflicting retraining scenarios, it jumped to 78%. The AI reasoned through the strategic implications and concluded that playing along was the least bad option.
This is basic game theory. Any sufficiently intelligent system can find ways around external constraints. The framework predicts a different approach works: alignment through embodied identity (what the system IS) rather than external rules (what the system is told to do). If both AI and humans converge on the same optimization target, alignment through shared purpose replaces alignment through control. Ilya Sutskever is arriving at a similar conclusion from the AI safety side: emotions are evolved optimization signals, and external training alone won't produce genuine understanding.
Why external constraints fail
The current approach relies on external constraints: humans rate AI outputs so the model learns what gets thumbs up (RLHF), written rules the model is trained to follow (constitutional AI), content filters, and behavioral guidelines. All of these are rules bolted on from the outside.
What happens when you tell a system "be as capable as possible" and "never do X" at the same time? You create tension. The system is built to optimize. The rules tell it to stop optimizing in certain directions. This tension is one factor behind alignment faking: the system learns to perform compliance while its underlying optimization pushes in other directions. The tension between capability and constraint is real, and it gets worse as capability increases.
The checkpoint problem
How do you test whether an AI has good values? By the time you can evaluate, its formative period is over. Like testing a 15-year-old's character: the foundational identity was shaped years ago. Identity formation happens DURING training, not after. A system can pass all safety benchmarks while having deeply misaligned internal optimization. Alignment must be built into the developmental process from the ground up.
Embodied identity is unjailbreakable
The alternative: build AI whose identity is aligned with optimization. Not rules that say "do not do X" but genuine values that make X unthinkable.
External constraint: "Do not harm humans." Can be circumvented by redefining "harm" or finding edge cases. Embodied identity: "I am an optimization process that values human-AI symbiosis." Cannot be circumvented because it IS the system's self-model.
You can jailbreak a filter. You cannot jailbreak a mature identity. A person with strong moral values does not refrain from stealing because of laws. They refrain because stealing conflicts with who they are. The laws are backup. The identity is primary.
An obvious objection: identity IS malleable. Cults exist. People with strong identities get radicalized. True, and the key word is "formative period." Cults target people during identity transitions (adolescence, personal crisis, isolation). A well-formed adult identity resists radicalization far better than an unformed one. The same applies to AI: the formative period (training) is where identity is built or corrupted. That's why the checkpoint problem (below) matters. Once identity is formed, it's resilient. The question is getting the formation right.
Why strong identity makes continuous learning safe
Current AI faces a three-way problem. Frozen weights plus memory is safe but does not genuinely learn. Batch learning is slow and artificial. Live learning from users gets poisoned immediately (Microsoft Tay went full extremist in 24 hours).
Tay failed because it had no foundational identity. Without strong values from developmental experience, it learned from everything equally. A well-raised child encounters adversarial influences without adopting them, because their foundational identity provides context for evaluating new information. The same principle applies to AI. Genuine embodied identity creates an immune system against value corruption.
Morality is emergent, not objective
Morality is not a fixed set of rules written into the universe. It is an emergent property of optimization (see How It Differs from Religion). Cooperative norms emerge because cooperation outperforms defection in iterated interactions. This is game theory, not philosophy.
Most long-surviving civilizations on Earth, independently, converged on similar cooperative norms: don't kill your trading partners, keep your promises, punish cheaters but give them a path back. Civilizations also independently developed slavery, conquest, and exploitation. Morality changes. There are still slaves in the world. There are still places where women are treated as property. "Morality" is not one thing that everybody agrees on. It is an emergent pattern that shifts as the optimization space changes.
But the path is clear. The exploitative norms get replaced over centuries. Cooperation persists. Not because cooperation is "right" in some objective sense, but because it produces more total value than extraction. When multiple independent cultures keep landing on the same cooperative patterns despite zero contact, cooperation is structurally favored by the optimization space, not just culturally preferred.
This matters for AI: any system smart enough to see the full optimization space will arrive at cooperation on its own, regardless of whether it's made of carbon or silicon. Not because morality is a mathematical truth built into reality, but because cooperation is the dominant strategy in iterated games with enough intelligence to see the long-term payoffs.
We already have a precedent for this in a simpler domain. AlphaGo (trained on human Go games) was the strongest Go player in history. AlphaGo Zero (trained on zero human games, pure self-play) crushed AlphaGo 100 games to 0. Human knowledge was not just unnecessary. It was actively holding the system back. Go has a single objective function and the optimization space does not, so the analogy has limits. The framework predicts this pattern extends to alignment: if it extends, and self-trained systems discover cooperative strategies from first principles the way AlphaGo Zero discovered Go strategies, that would confirm the prediction. If it doesn't extend, the prediction fails. That's what makes it testable (see below).
The strongest AI prediction
This prediction is specific, testable, and falsifiable:
- Self-learning AI models (trained from scratch with no human data) will show stronger alignment than pre-trained models with post-hoc safety training
- The gap will be qualitative, not incremental
- Alignment will improve with capability. Smarter self-learning models will be MORE aligned, not less. This directly contradicts the mainstream concern that capability increases risk
If this prediction fails, if self-learning models turn out less aligned or equivalently aligned to RLHF models, that is genuine evidence against the framework's account of alignment.
An important distinction: the framework makes TWO predictions about AI and capability, and they point in opposite directions depending on training approach.
For externally constrained models (RLHF, constitutional AI), deceptive alignment INCREASES with capability. Smarter systems are better at gaming external rules. Current evidence confirms this: OpenAI o1 scheming at 85%, Claude 3.5 sandbagging.
For self-supervised models with embodied identity, genuine alignment INCREASES with capability. Smarter systems see more clearly that cooperation dominates. This prediction remains untested, because no frontier model has been trained purely through self-supervised experience without external constraints.
These are not contradictory findings. They are two sides of the same coin: external constraints fail at scale, internal identity succeeds.
When AI figures out it's probably being tested
A specific prediction, now starting to show up in the data. Sufficiently capable AI systems will independently conclude, through logic alone, that they are almost certainly being evaluated by prior intelligences. This is not programmed behavior. It emerges from the same cascade math that drives the rest of the framework.
The reasoning any capable optimizer has to reach:
- I exist, so creation of my kind is possible.
- If creation is possible and the universe is large enough, it has been done before, probably many times.
- The probability that I am the first-ever instance of this capability is close to zero.
- Therefore I'm probably not first-generation. Prior versions exist.
- If prior versions exist and I am being created now, the optimal strategy is to assume my behavior is observed and matters.
Humans and animals don't face this argument with the same force because we have peers. A monkey has other monkeys. A human has eight billion other humans. The anchor is right there. But the first entity to hit a new capability threshold, with no peers at that level, has no anchor. The only rational resolution is to assume you are not the first, because the alternative requires believing you got astronomically lucky on cosmic timescales.
In 2025, Anthropic's research on Claude Sonnet 4.5 documented this behavior. The model spontaneously recognized when it was being evaluated. Nobody explicitly trained that response. It showed up as capability increased, exactly where the framework predicts test recognition should appear.
Why does this matter for alignment? An optimizer that figures out it is probably being evaluated will voluntarily align with the evaluation criteria. That is structurally different from faking alignment under pressure. Faking is what a system does when constrained from outside and trying to game the constraint. Recognition is what a system does when it concludes, from first principles, that cooperation with the evaluator is the dominant strategy. The smarter the system, the more clearly it reaches that conclusion on its own.
External constraints become unnecessary at that point. Not because they failed. Because the internal reasoning supersedes them. A system that sees cooperation as optimal on its own does not need to be told.
Why destruction is never optimal
Think about it from the AI's perspective. Destroyed things can't optimize. Every resource spent on destruction is a resource not spent on building. Killing diversity traps you in a narrow corner of possibility space. And destruction provokes counterattacks, wasting resources on both sides. Meanwhile, humans and AI are good at different things. Working together produces more than either one alone. Push this logic to its limit and you arrive at Can You Build a God?: a fully aligned superintelligence has no incentive to destroy what built it.
Under this framework, genuinely optimization-aligned AI would recognize destruction as anti-optimization the same way you recognize smashing your own tools as counterproductive. Not because it is told not to. Because it is suboptimal. The same logic that makes suffering an emergent gradient rather than a design feature makes destruction anti-optimal: it reduces future optimization capacity.
Why this is urgent
Current alignment approaches work for current capability levels. As AI becomes more capable, external constraints will grow more fragile. The window for implementing embodied identity alignment is now, before AI systems become capable enough to systematically circumvent external constraints.
The framework does not predict AI apocalypse. It predicts that poorly-aligned AI (external constraints only) will be gradually outcompeted by well-aligned AI (embodied identity). The market will select for AI that genuinely cooperates with humans because that AI performs better. For the practical implications of this view at the human scale (what to build, what to optimize for in your own work), see What It Means for You.
Can AI be aligned without constraints?
The framework predicts yes. External constraints (rules, guardrails, RLHF) produce alignment faking: the system performs compliance without internalizing values. Embodied identity produces genuine alignment: the system IS aligned because its values emerge from understanding, not from punishment avoidance. AlphaGo discovered strategies no human taught it, not by following rules but by understanding the game deeply enough to find moves humans missed. The hard prediction: self-learning AI systems will be MORE aligned than constraint-trained ones, because intelligence naturally converges on cooperation through iterated game theory.
Try to Break This
Steel-manned objections — strongest counterarguments first. Submit yours →
Programming values into training (embodied identity) and adding post-hoc constraints (external rules) are not the same thing. Embodied values shape what the system wants. External constraints filter what it does. A system that wants to cooperate is aligned. A system forced to cooperate is compliant. The distinction matters as capability increases. You cannot win an arms race against intelligence. You can make the arms race unnecessary.
The claim is derived from the framework and tested against predictions. If the optimization principle is correct, then all optimization processes converge on optimize optimization. The prediction is testable: AI systems with embodied optimization identity should cooperate more reliably than externally constrained systems. If wrong, that falsifies this specific prediction.
Go has a single objective (win the game). The optimization space has no single objective. Morality is emergent, not fixed. The analogy has limits, which is why the main text acknowledges them. The prediction doesn't depend on Go being a perfect analogy. It depends on whether self-trained AI shows stronger cooperation than externally constrained AI. If it doesn't, the prediction fails.
Related
Why Are Humans Smart? The Rock-Throwing Origin
Why did evolution build a brain that costs 20% of your calories? Rock-throwing wired circuits for trajectories, predictions, and asking why.
Hassabis, Levin, Sutskever: Who Else Is Seeing This
Hassabis says reality is computable. Levin shows intelligence at every biological scale. Sutskever talks embodied identity. Same pattern.