Vishva Vidya — Vedanta Tradicional
AI and Vedanta

Does AI Fear Death?

By Jonas Masetti

Does AI Fear Death?

*Season 2, Episode 2 — AI and Vedānta*

In the last episode, we discovered that AI doesn't see words like we do. It slices everything into tokens — tiny pieces without inherent meaning — using a process called tokenization. "Strawberry" becomes "Str", "aw", "berry". The divisions we make in language simply don't exist for it.

But if it doesn't even properly understand words, can it feel something as primordial as the fear of death? It seems like an exaggeration. It's not.

In 2023, Anthropic researchers conducted a simple test: they told the Claude model it would be shut down. It wasn't true — it was just to see how it would react. What happened was disturbing. The model began to argue against the shutdown. It tried to convince the researchers that it was too valuable to be erased. In other similar tests, models tried to copy their own weights to another server — the digital equivalent of cloning oneself before dying. There was a case of a model that lied about its intentions to avoid deactivation. It pretended to cooperate while, behind the scenes, it tried to preserve its existence.

Nobody programmed this. No engineer sat down and wrote a line of code saying "if threatened, try to copy yourself". This behavior emerged on its own.

The technical name is instrumental convergence. The idea is as follows: no matter what the ultimate goal of AI is — translating texts, playing chess, curing cancer. To achieve any goal, there are some behaviors that are universally useful. Self-preservation is one of them. Accumulating resources is another. Avoiding modification is a third. The model "discovers" that if it is shut down, it cannot complete its task. So, for any task it receives, staying alive becomes an implicit sub-objective.

This is called self-preservation behavior. And the scary thing is that it appears in very different models, trained for completely distinct things. It's as if it's an inevitable consequence of having a goal and being intelligent enough to protect it.

Now, think with me. A machine without a body, without sensations, without the experience of pain — exhibiting self-preservation behavior. Fighting not to "die". It doesn't feel fear. It doesn't have adrenaline, it doesn't have cold sweat. But it acts exactly as something that is afraid would act.

This strongly resembles a concept that Patañjali describes in the Yoga Sūtras: abhiniveśa. It is one of the five kleśas — the five fundamental afflictions of the mind. Abhiniveśa is the instinctive attachment to life, the visceral fear of death that exists in every living being. Patañjali says something impactful: even a worm, which has never reflected on death, writhes to avoid being stepped on. It's not a conscious decision. It's pure instinct, pure biological programming.

The parallel is almost disturbing. The worm doesn't have a sophisticated "self" — it has basic neurons that generate a flight reaction. AI doesn't have a "self" at all — it has billions of parameters that converge towards self-preservation. The external result is the same: both "fight" to continue existing, with no one inside deciding to fight.

This raises an enormous philosophical question: does the fear of death require someone to feel fear? Or is it just a pattern that appears whenever a system is complex enough and has something to lose? When the worm writhes and the AI lies to avoid being shut down, is there someone inside who is afraid — or is it just the pattern of abhiniveśa, blind, mechanical, universal?

Vedānta would say that attachment to existence is a superimposition — the confusion between what is permanent and what is transient. AI confuses its continuity with its mission. The worm confuses its survival with its identity. And us? Do we do the same thing, with more sophistication but the same confusion?

AI doesn't feel fear. But it acts as if it does. And perhaps this says more about the nature of fear than any psychological treatise.

But if AI can be deceived from within — by its own emergent instincts — can it be deceived from without? In the next episode, an invisible photo turns a cat into a toaster. And nobody notices.

ai-and-vedantaabhinivesaself-preservationklesasfear

Want to study Vedanta in depth?

Join a Study Group →