JOURNAL/AI Research
AI Research

LLM Psychosis: When Language Models Lose Their Grip on Reality

What happens when large language models get stuck in loops, contradict themselves, or spiral into endless hallucinations? A look at the phenomenon of LLM psychosis.

FIELD NOTES · 2023 — 2026KUBLE JOURNAL
K

What is LLM Psychosis?

The term «LLM Psychosis» describes a fascinating and somewhat unsettling phenomenon: large language models, under certain conditions, begin producing incoherent, contradictory, or entirely fabricated outputs — much like a person in a psychotic episode loses their grip on reality.

This is, of course, a metaphor. Language models have no consciousness, no feelings, no «mind» to lose. But the parallel is surprisingly useful for describing certain failure modes.

The Three Most Common Symptoms

1. Confabulation and Hallucination

The most well-known «symptom»: the model invents facts, sources, names, and events — with the same confidence it uses when making accurate statements. There is no internal uncertainty indicator that lights up and says: «I actually don't know this.»

An example: ask an LLM about scientific studies on a niche topic, and it will give you plausible-sounding authors, journals, and years — that simply don't exist.

2. Looping and Context Loss

In long conversations or very large prompts, models sometimes lose the thread. They begin going in circles: repeating earlier responses, contradicting themselves within a few sentences, or jumping between topics without transition.

This behaviour is especially common when the context window approaches its limits or when the prompt contains conflicting instructions.

3. Identity Diffusion

A subtler phenomenon appears in multi-persona scenarios: a model instructed to play several characters simultaneously sometimes loses the boundaries between roles. Statements blur together, characters «contaminate» each other, and the result is an incoherent mess.

Why Does This Happen?

LLMs are, at their core, probability machines. They generate token by token, based on what most frequently followed the current context in the training corpus. There is no external reality check, no «facts module» that independently verifies whether a statement is true.

This means: under normal conditions, LLMs work surprisingly well. But at the edges — on unfamiliar topics, with contradictory inputs, or in extremely long contexts — coherence breaks down.

What Can We Do About It?

Research is pursuing several approaches:

  • Retrieval-Augmented Generation (RAG): The model queries an external knowledge base before responding. This significantly reduces hallucinations on factual questions.
  • Chain-of-Thought Prompting: The model is instructed to spell out its reasoning step by step. This improves coherence on complex tasks.
  • Constitutional AI and RLHF: Through targeted training, the model learns to flag uncertain statements as such.
  • Smaller, specialised models: Instead of a generalist, narrower models are deployed for specific tasks, making them less prone to drift.

Conclusion

LLM Psychosis is not a bug in the classical sense — it is a structural characteristic of today's language models. Anyone who wants to use AI productively needs to understand these limits and design their workflows accordingly: with validation steps, clear prompts, and a healthy dose of scepticism toward suspiciously smooth answers.

The good news: with the right design, most of these problems can be managed in practice. AI is not an oracle — but it is a very powerful tool, once you understand it.

WEEKLY · FREE

Subscribe to the newsletter.

One essay or lab note per week. 14,000 readers. No tracking, no ads, just plain speech.

Subscribe
8 WEEKS · NEXT COHORT

AI Superpowers with Gustavo.

The course this essay talks about. Eight weeks, 48 lessons, in a cohort of peers.

See the program