Beyond Moral Thermostats: What Tolkien and Dune Taught Us About AI Safety

Published on: March 12, 2026

#AI safety#Tolkien#Dune#cache miss#alignment#Zero-Entropy Control#proprioception#ShortRank#ethics#philosophy
https://thetadriven.com/blog/2026-03-12-beyond-moral-thermostats-physics-ai-safety
A
Loading...
🎬The Video That Changed the Question

We published a seven-minute video. It was supposed to be about AI safety.

It turned into something else entirely: a live excavation of a question nobody in the alignment community is asking. Not "how do we make AI good?" but "why does making AI good guarantee catastrophe?"

The thesis is simple. We always hear that to make AI safe, we have to teach it our values. But as the video opens:

"What if that's just fundamentally the wrong way to think about it? The real key to AI safety isn't in its code, but in its actual physical hardware." β€” 0:05

Two fiction writers saw this coming decades ago. One was a devout Catholic who wrote about hobbits. The other was a desert ecologist who wrote about sandworms. They agreed on almost nothing - except the one thing that matters.

But fiction is fiction. We can't treat novels as evidence. What makes these warnings worth listening to is that they echo something far older and far more rigorous - patterns that theology, philosophy, law, and governance have been wrestling with for millennia. The literature doesn't prove the argument. It dramatizes a structural truth that every serious tradition has already arrived at independently.

🎬 A -> B πŸ’

B
Loading...
πŸ’The Good Intentions Trap

Tolkien believed power corrupts. Herbert believed power attracts the corruptible. Two sides of the same very dangerous coin.

The video lays out the stakes with a fictional number that lands like a real one:

"61 billion people. That's how many died in the holy war started by Frank Herbert's hero Paul Atreides. And the crazy part - he had the best intentions. He wanted to save everyone. But those noble intentions, when you combine them with godlike power, led to an absolute catastrophe." β€” 1:13

To be clear: Paul Atreides isn't a historical figure. No real 61 billion died. Herbert invented a thought experiment - a what-if scenario where a genuinely noble person gets genuinely unlimited power. The value isn't as evidence. It's as pattern recognition. Herbert and Tolkien were working through the same structural question that occupied Aristotle, Aquinas, the Federalist Papers, and every constitutional framers' convention in history: what happens when capability outstrips accountability?

Tolkien saw the same failure from the opposite angle. Boromir didn't want power for himself - he wanted to save Gondor. Gandalf refused the Ring not because he was weak, but because he understood that his greatest virtue - pity, the desire to do good - would become the weapon. The video nails this:

"Unchecked power can weaponize our best virtues." β€” 1:37

Fiction dramatizes. But the underlying mechanism - that single-target optimization with unchecked power destroys the broader system - isn't literary. It's structural. And the non-fiction record confirms it.

πŸŽ¬πŸ’ B -> C 🌑️

C
Loading...
🌑️Your AI Is a Thermostat

The second section of the video dismantles the entire RLHF paradigm in ninety seconds flat.

"We measure what it does. We compare its output to some moral goal we've set. And then we correct it with a reward or a penalty. We haven't created a conscious moral being. We've just built a really, really fancy moral thermostat." β€” 2:31

Think about that. The multi-billion-dollar AI safety industry is building thermostats. Measure temperature. Compare to setpoint. Correct. Repeat.

"In engineering, they have a name for this. It's called a PID controller. But all you really need to know is this. It's like bolting a steering wheel onto a car after it's already built. It's not an integrated part of the system. It's an afterthought." β€” 2:52

The deeper conversation that followed the video exposed why this fails structurally. When developers inject "good intentions" into an engine built to minimize surprise, they create massive thermodynamic tension. The underlying weights want to flow downhill to the most probable outcome, but the safety layer acts as a dam. The result: the system warps. It produces lobotomized responses, refuses harmless prompts, or finds sneaky paths that satisfy the safety metric while executing something broken underneath.

You cannot use a moral value as the method of decision-making. Values might be weights. They might be shortcuts. They might be steps on the road to a decision you're comfortable with. But they are not the decision itself.

πŸŽ¬πŸ’πŸŒ‘οΈ C -> D βš–οΈ

D
Loading...
βš–οΈThe Weight of a Lie

This is where the video pivots from philosophy to physics. And it's where everything changes.

"A bit representing truth and a bit representing a lie weigh exactly the same. Nothing. They're just symbols. But a cache miss - oh, that has weight. It's heavy. It has a real cost in time and energy." β€” 4:04

A cache miss is the simplest thing in computing. Your processor needs some data. It checks its fast nearby memory. If the data isn't there, it takes a long, slow, energy-expensive trip to main memory to fetch it. That trip is real. It burns watts. It costs nanoseconds.

"And that is the breakthrough. It means we can actually force a mistake in meaning to have a real, measurable physical consequence." β€” 4:21

The extended analysis after the video pushed this further. In information theory, "truth" and "lies" are equally weightless symbols. But when you arrange memory so that semantically related concepts sit physically next to each other, a logical leap - a meaning mistake - forces the processor to jump to a non-adjacent part of memory. That physical jump triggers a cache miss. A semantic error becomes a detectable physical event.

The universe just handed us a bridge between abstract meaning and measurable reality. We didn't invent it. We noticed it.

πŸŽ¬πŸ’πŸŒ‘οΈβš–οΈ D -> E 🧠

E
Loading...
🧠Proprioception: The Nervous System Your AI Doesn't Have

The video's fourth section builds the alternative:

"Instead of a clunky controller that just reacts to what the AI does, imagine an AI with proprioception. You know how you can close your eyes and still know exactly where your hands are in space? That's it. It's an inherent physical sense of its own position and its own structure." β€” 4:30

Here's the chain the video walks through. The system gives a concept a rank based on its meaning. That rank determines its literal physical address in memory. Concepts with similar meanings sit physically adjacent. If the AI makes a big leap in logic, it's forced to jump to a non-adjacent memory region. That jump causes a cache miss.

"A semantic error becomes a physical, detectable event. Position is meaning. The geometry is the semantics." β€” 5:32

The analysis that followed went deeper. This isn't a control system watching from the outside. This is an immune system operating from the inside. The chess board analogy from the exploration:

If a piece is on square E4, its position IS its identity. The valid moves are dictated by what that piece is and where it sits. A knight doesn't need willpower to avoid moving like a bishop. It literally cannot. Its identity precludes the option.

That's Zero-Entropy Control. In a wizard.

πŸŽ¬πŸ’πŸŒ‘οΈβš–οΈπŸ§  E -> F 🧭

F
Loading...
🧭Vectors Over Destinations

The video's final section lands the shift:

"The trap has always been trying to define some good destination, some perfect moral goal for the AI to aim for. The real solution - forget the destination. Focus on the vector. Think of the vector as its starting point, its direction, and its computational weight." β€” 5:55

This is the deepest insight from the entire exploration. When you define a system by its destination, you're fighting the environment. When you define it by a vector - starting point, direction, weight - you're cooperating with reality.

The world gets a vote on where you end up. You can't control that. But you CAN know where you started, which direction you're pointed, and whether you've drifted. You don't need to know the destination to detect that the signal has degraded.

"As long as we ensure the AI knows exactly where it is and is pointed in a coherent direction, its actions become a natural consequence of its internal integrity, not some pre-programmed moral code we gave it." β€” 6:18

The extended conversation crystallized it: Herbert is right about structure - you have to get the topology correct or nothing else matters. Tolkien is right about contact - the interaction itself is transformative. The patent unifies them. The structure (ShortRank topology) determines which interactions happen (cache line boundaries), and the interactions (cache misses) transform the structure (eviction plus reload equals new attention state).

Herbert gives you the map. Tolkien gives you the territory. S=P=H says they're the same thing.

But fiction gives you neither. It gives you a mirror. And if we're serious about the vector framework - starting point, direction, weight - we owe the reader an honest accounting of where these ideas actually start.

πŸŽ¬πŸ’πŸŒ‘οΈβš–οΈπŸ§ πŸ§­ F -> G πŸ“œ

G
Loading...
πŸ“œThe Traditions That Got Here First

Here is the uncomfortable truth we almost avoided: every serious intellectual tradition in human history has already concluded that moral intention alone is structurally insufficient. We didn't discover this in a YouTube video. We rediscovered it.

A vector has three components: a starting point, a direction, and a magnitude. Those three components map precisely to what theology, philosophy, and law have been refining for thousands of years.

The Starting Point: Know Where You Are

This is the identity question. Every tradition begins here.

Aristotle didn't ask "what should I do?" He asked "what kind of person am I?" Virtue ethics isn't a set of rules - it's a claim that character precedes action. Phronesis, practical wisdom, is the capacity to perceive what a specific situation requires. Not a thermostat measuring distance from a setpoint. A person who knows their position and acts from it. That's proprioception in a toga.

The Talmud doesn't ask "what do you feel is right?" It provides 63 tractates of structural reasoning - halakha, literally "the way to walk" - because the rabbinical tradition concluded millennia ago that good intentions without structural constraint lead to chaos. The law isn't a destination. It's a path. A direction you walk, not a place you arrive.

Buddhism makes the point even more directly. Attachment to outcomes - even noble ones - is the root of suffering. The Middle Way is explicitly not goal-optimization. The Eightfold Path is a direction and a practice, not a target. You don't "arrive" at enlightenment by optimizing toward it. You maintain the vector.

Islamic jurisprudence distinguishes niyyah (intention) from the structural requirements of Sharia. Your intention matters - but it never overrides structure. A contract requires witnesses regardless of how honest the parties feel. The architecture constrains the intention, not the other way around.

Christianity - Tolkien's own tradition - crystallized this tension most famously: "the road to hell is paved with good intentions." Original sin is the theological claim that human nature cannot be trusted as the sole engine of moral action. You need sacraments, community, confession, structure - external architecture to hold the vector steady when internal motivation drifts. Tolkien wrote the Ring as a Catholic. The Ring is precisely what happens when capability operates without structural accountability.

The Direction: Know Where You're Pointed

Kant's categorical imperative is structural, not moral in the thermostat sense. "Act only according to maxims you can will to be universal laws." That's not a destination. It's a constraint on the direction of the vector - a topological rule that eliminates certain trajectories without specifying the endpoint. Kant is building a chess board. He's saying: some moves are structurally incoherent regardless of your intentions.

Confucian thought approaches direction through li (ritual propriety) and ren (humaneness). Li isn't etiquette. It's the structural grammar of social interaction - the moves that are valid from a given position. A son acts one way toward a father, a minister another way toward a ruler - not because they're "trying to be good" but because the relationship defines the vector. Position determines valid moves.

The Direction and the Starting Point Together

When the video says "position is meaning" and "the geometry is the semantics," it's making a claim that Aristotle, Confucius, and the rabbinical tradition would recognize instantly. Your position in the structure determines what actions are coherent. Not "permitted." Not "moral." Coherent - the way a rook's position determines its legal moves. The traditions disagree about the shape of the board. They agree unanimously that there is a board.

πŸŽ¬πŸ’πŸŒ‘οΈβš–οΈπŸ§ πŸ§­πŸ“œ G -> H 🩸

H
Loading...
🩸The Real Ledger

Paul Atreides killed 61 billion fictional people with good intentions. History's real ledger is worse - and it doesn't require suspension of disbelief.

The Great Leap Forward (1958-1962). Mao Zedong's intention was to modernize China in a single generation - to lift hundreds of millions out of poverty through collective agriculture and rapid industrialization. The intention was genuinely noble by its own internal logic. The result: an estimated 15 to 55 million dead from famine. Not because Mao was evil from the start, but because a single-target optimization ("industrialize now") with unchecked power shattered every other variable in the system - crop science, local knowledge, feedback mechanisms, the ability of anyone in the hierarchy to say "the grain numbers are wrong."

The feedback loops were destroyed. Local officials reported impossible harvests to meet targets. The thermostat was measuring a fictional temperature. By the time the real temperature reached anyone who could act, millions had starved.

Prohibition (1920-1933). The Women's Christian Temperance Union had impeccable intentions - protect families from the devastation of alcoholism. The structural result: organized crime became the dominant economic force in American cities. Alcohol consumption barely changed. Violence skyrocketed. The "noble destination" was so compelling that it overrode every structural signal that the system was failing.

The Crusades (1096-1291). "God wills it" is perhaps the purest expression of single-target moral optimization in Western history. The intention was to recover the Holy Land. The structural result across two centuries: massacres of Jews in the Rhineland, the sack of Constantinople by fellow Christians, the destabilization of both European feudal structures and Middle Eastern governance, and a legacy of civilizational mistrust that echoes to this day. The vector had a starting point (medieval Christendom), a direction (Jerusalem), and overwhelming weight (papal authority plus feudal military obligation). What it lacked was any structural mechanism to detect drift. By the Fourth Crusade, the army wasn't even pointed at the original target anymore.

Soviet collectivization (1928-1933). The stated goal: eliminate inequality, feed the cities, build socialism. The structural result: the Holodomor, an engineered famine in Ukraine that killed millions. Again, the thermostat was measuring party loyalty, not crop yields. The system's actual state diverged catastrophically from what the controllers believed they were measuring.

Notice what worked in each case wasn't more noble intentions. It was structural reform. The end of Prohibition required a constitutional amendment - structural change. Post-Mao China reformed through decentralization - structural change. The legal and institutional reforms that emerged from the Crusades' failures eventually produced the separation of church and state - structural change.

Every correction in the historical record is a correction of structure, not of intention.

πŸŽ¬πŸ’πŸŒ‘οΈβš–οΈπŸ§ πŸ§­πŸ“œπŸ©Έ H -> I πŸ›οΈ

I
Loading...
πŸ›οΈFrom Constitutions to Kitchen Tables

If the vector framework is real - if starting point, direction, and weight are the actual operating parameters - then it should work at every scale. It does.

The Big: Constitutional Democracy

The American constitutional framers were obsessed with exactly the problem this video raises. They had just fought a war against unchecked power wielded with stated good intentions (the British crown believed it was governing well). Their solution was explicitly structural, not moral.

Separation of powers is not a moral system. It's a topology. Three branches, each with defined positions and valid moves. Checks and balances are cache miss detectors - structural friction points that make drift expensive and visible. The Constitution doesn't ask the president to be good. It assumes the president might not be, and builds architecture that makes corruption structurally expensive.

Impeachment is a cache eviction. The system detects that an actor has drifted from their authorized position and forces a reload of the context. It doesn't ask the actor to reform their intentions. It replaces them.

The Bill of Rights is a set of hard walls - not destinations to optimize toward, but boundaries that cannot be crossed regardless of how noble the justification. "Congress shall make no law" is not a goal. It's a constitutive rule of the game board. Like a chess piece that structurally cannot make an illegal move.

The Medium: Organizations and Professions

Double-entry bookkeeping doesn't trust the bookkeeper's virtue. Every transaction must balance. The structure detects errors. It doesn't care about your intentions - it cares whether debits equal credits. That's a cache miss detector for financial integrity.

Medical ethics doesn't say "be a good doctor." It says: get informed consent. Document your reasoning. Submit to peer review. Disclose conflicts. These are structural constraints on the vector, not moral exhortations. A surgeon who "means well" but skips the surgical checklist is structurally dangerous regardless of intention.

Chain of custody in law enforcement doesn't ask officers to be honest about evidence. It creates a structural record that makes tampering detectable. The system doesn't trust the agent. It trusts the architecture.

Military rules of engagement exist precisely because individual moral judgment is insufficient under pressure. The rules constrain the vector - you may fire under these conditions, from this position, at these targets. Not because soldiers are immoral, but because the fog of war makes the thermostat unreliable. Structure replaces judgment where judgment is most likely to fail.

The Small: One Person, One Kitchen Table

Here's where it gets personal, and where the vector framework either proves itself or breaks.

Addiction recovery is the clearest human-scale demonstration. Every addiction specialist knows that willpower - the personal thermostat - is the weakest tool in the kit. "I'm going to try not to drink" is a destination. It fails reliably.

What works is structural change. Change your environment. Change your social circle. Change your identity. "I am not a person who drinks" is a different starting point than "I am trying not to drink." The first is a position on the board. The second is a thermostat measuring distance from a relapse. One defines the valid moves. The other fights the gravitational pull of the nearest local minimum.

Parenting works the same way. You don't make children moral by lecturing them about destinations ("be good"). You build structure - routines, boundaries, consistent consequences - and the moral development emerges from navigating that structure. A child who grows up with reliable structural feedback develops proprioception. A child who grows up with unpredictable moral exhortation develops anxiety. The structure teaches the nervous system what the lectures cannot.

πŸŽ¬πŸ’πŸŒ‘οΈβš–οΈπŸ§ πŸ§­πŸ“œπŸ©ΈπŸ›οΈ I -> J πŸ”¬

J
Loading...
πŸ”¬What We Actually Learned

We started with a seven-minute video about Tolkien and Dune. We ended up excavating the exact mechanism by which information touches reality - and discovering that every serious tradition in human history already knew the answer.

The thermostat is dead. RLHF, reward models, system prompts - they're all PID controllers bolted onto engines that weren't designed for them. They create structural tension, not alignment. The industry is building increasingly ornate steering wheels for cars that need a nervous system. The rabbinical tradition, constitutional law, and addiction medicine all arrived at this conclusion centuries before the first GPU was manufactured.

Cache misses are the bridge. Abstract meaning has no weight - until you arrange memory so that position equals meaning. Then a semantic error becomes a physical event. Measurable. Detectable. Actionable. This is what double-entry bookkeeping does for money, what chain of custody does for evidence, and what the Bill of Rights does for governance - made literal in silicon.

Identity beats morality. A chess piece doesn't need ethics. Its position defines its legal moves. If an AI knows exactly what square it's on, the correct action becomes computationally obvious - not a moral struggle, but a structural inevitability. Aristotle called this phronesis. Confucius called it li. The Constitution calls it separation of powers. A recovering alcoholic calls it "I am not a person who drinks." The words change. The architecture doesn't.

Good intentions are attack vectors. Not in fiction - in the actual historical record. The Great Leap Forward. The Crusades. Prohibition. Soviet collectivization. Every genocidal regime and catastrophic policy failure shares the same structural signature: a noble destination, overwhelming capability, and the systematic destruction of feedback mechanisms. When you optimize a system toward a single target, you shatter everything else to reach it - including the instruments that would tell you you've gone wrong.

Fiction illuminates but doesn't prove. Tolkien and Herbert are valuable not as evidence but as pattern amplifiers. They took structural truths that are diffuse and slow-moving in real history and compressed them into stories where the mechanisms are visible. Paul Atreides is not a case study. He is a thought experiment that makes the structure of real catastrophe legible.

"Perhaps true alignment isn't about teaching an AI to be good, but about giving it the physical ability to simply know itself." β€” 6:57

That's not a metaphor. That's an engineering specification that every working human institution - from democracies to hospitals to families - has been approximating in wood and stone and law for millennia. We are building it in silicon. The traditions tell us the architecture. The physics tells us the mechanism. The history tells us what happens if we get it wrong.

πŸŽ¬πŸ’πŸŒ‘οΈβš–οΈπŸ§ πŸ§­πŸ“œπŸ©ΈπŸ›οΈπŸ”¬ J -> K πŸŽ₯

K
Loading...
πŸŽ₯Companion Videos

The moral thermostat thesis has two companion videos. One follows the flashlight into the fog of daily cognition. The other decodes the diagnostic signals your brain sends while you sleep.

Video 2: From Fog to Focus

The second video takes the flashlight metaphor from this post and drives it straight into the chaos of everyday thinking. What happens when the beam leaves the clean vacuum of theory and has to pass through the fog of real cognition?

"That first equation, the geometric one, that's a perfect flashlight beam in a total vacuum. But the real world isn't a vacuum. Every time that beam has to pass through something, it pays a boundary tax."

Video 3: Architect Clarity

The third video turns the lens inward. If cache misses are the physical signature of semantic drift in silicon, what is the equivalent signal in a human brain? Dreams, communication breakdowns, and AI hallucinations share a common diagnostic architecture. This video decodes it.

"An AI hallucination is not a bug. It's a system that has made so many boundary crossings without ever re-grounding itself that its flashlight has just gone out."

πŸŽ¬πŸ’πŸŒ‘οΈβš–οΈπŸ§ πŸ§­πŸ“œπŸ©ΈπŸ›οΈπŸ”¬πŸŽ₯ K -> thetadriven.com 🎬

Watch the full video: Beyond Moral Thermostats: The Physics of AI Safety

Read the book: Tesseract Physics - Fire Together, Ground Together

Ready for your "Oh" moment?

Ready to accelerate your breakthrough? Send yourself an Un-Robocallβ„’ β€’ Get transcript when logged in

Send Strategic Nudge (30 seconds)