9 min read

The New Evidence That AI Reasoning Is Mostly Theater

A language model lays out a neat sequence of steps, reaches the right answer, and sounds uncannily deliberate. We watch that performance and instinctively map it onto our own experience of thinking. The resemblance is strong enough that many people, including plenty of people building these systems, started to talk as if the resemblance settled the question.

A recent line of research cuts straight into that assumption. Its claim is not that models are useless, or that every impressive answer is fake. The claim is narrower and more unsettling: much of what we call “reasoning” in large language models may be a polished form of retrieval and pattern completion, one that breaks when the surface pattern shifts.

That matters because the industry has built a lot on top of the opposite belief. Product strategy, safety claims, benchmark culture, and user trust all lean on the idea that step-by-step output reveals a system manipulating abstract rules in a robust way. If that story is wrong, even partly, a lot of confidence starts to look cosmetic.

A cleaner way to ask the question

The core problem with evaluating model reasoning is contamination. Real-world tasks are messy. Training data is enormous. When a model answers a puzzle, you rarely know whether it inferred a rule or recognized a familiar structure from somewhere in its data.

So the researchers built a sandbox designed to strip away that ambiguity. They call it DATAALCHEMY. The ingredients are intentionally simple: letters instead of atoms, strings instead of molecules, transformations instead of chemistry. A task might ask the model to apply operations to a sequence like APPLE, such as shifting letters forward in the alphabet or composing multiple transformations in order.

That simplicity is the point. In a controlled world, success means more. Failure also means more. There is nowhere for vague “world knowledge” to hide, and fewer chances for the model to bluff its way through with statistical familiarity.

If a system really grasps an underlying rule, this kind of synthetic domain should not be especially threatening. Humans often generalize better in stripped-down environments because the principle stands out. A child who understands what addition is can add apples, blocks, or dots on paper. The wrapping changes. The operation survives.

Competence that does not travel

In the controlled setting, the models often performed well on the exact kinds of examples they saw during training. That part is not shocking. Pattern learners are supposed to be good at familiar territory.

The trouble began when the researchers moved only slightly beyond that territory. They tested new combinations of already known transformations. Performance dropped sharply. The system could execute operation A and operation B in contexts it had memorized, yet stumbled when asked to compose A and B in a configuration it had not seen before.

That is a revealing kind of failure. It suggests the model is not carrying around a compact, reusable rule in the way we usually mean by understanding. It is carrying a set of learned associations tied closely to presentation and distribution.

The same pattern showed up when the researchers changed the “elements” being transformed. Train on sequences using one subset of letters, then test on sequences built from another subset. Again, accuracy collapsed. The model had not learned a general procedure that transfers across tokens. It had learned something narrower and more local.

Length was another fault line. A model trained on shorter sequences struggled with longer ones, even when the underlying operation stayed the same. That does not prove a total absence of abstraction. But it does tell you the abstraction, if present, is weak enough to snap under modest extension.

The most sobering result may be what happened next. A small amount of fine-tuning on these “new” cases often restored performance. On paper, that can look encouraging. In practice, it reinforces the concern. The model did not uncover a deep principle and then apply it broadly. It absorbed another patch of the pattern landscape.

Chain-of-thought may be output, not introspection

People often treat chain-of-thought as if it were a window into the machine’s mind. The model writes down intermediate steps, therefore those steps must reflect an internal reasoning process. That leap was always tempting, and always a little suspect.

The new evidence pushes hard against it. A reasoning trace can be useful without being faithful. It can help the model stay on track, much as writing on scratch paper helps a person reduce errors. But usefulness is not the same as transparency. The visible explanation may be more like a polished receipt than a live feed from cognition.

This is one reason chain-of-thought is so easy to overinterpret. Language is a performance medium. If the sequence of words looks orderly, we grant it a kind of legitimacy. We are primed for that. Humans infer competence from fluency all the time. A confident speaker with a whiteboard marker can get remarkably far before anyone notices the equations are decorative.

Large models exploit that bias without intending to. They are extremely good at producing the shape of explanation. They know what a careful derivation sounds like, how a legal analysis is usually staged, where a proof tends to insert a lemma, when a consultant would probably say “therefore.” Form and function line up often enough that the illusion holds.

Until the format shifts.

The fragility of wording is the tell

One of the study’s most important implications is not about synthetic chemistry at all. It is about prompt sensitivity. If a system truly grasps a rule, small changes in phrasing should usually leave the underlying capability intact. The wording may matter at the margins, but the structure should survive.

With language models, it often does not. Reorder the request. Add a seemingly irrelevant sentence. Swap a term for a synonym. Change the output format. Systems that looked methodical can veer into nonsense while preserving the posture of logic.

That combination is what makes the problem dangerous. Wrong answers are easy to spot when they are chaotic. Wrong answers wrapped in elegant scaffolding are much harder. The trace feels earned. The intermediate steps create a sense of auditability even when the audit trail is mostly decorative.

This helps explain a strange everyday experience with chatbots. You ask the same substantive question twice in slightly different ways and get two incompatible chains of reasoning, each delivered with equal poise. Users often read that as inconsistency at the edges. In many cases it is more revealing than that. It shows the “reasoning” is coupled tightly to linguistic presentation, not just to task structure.

Why this fooled so many smart people

Part of the answer is simple: these systems are genuinely impressive. A model that can summarize a legal brief, fix a bug, explain Bayes’ rule, and draft a decent email in the same afternoon deserves some awe. People are not irrational for being struck by it.

Another part is that benchmark culture rewards the appearance of generality. If a model scores well across enough reasoning tasks, it is natural to describe the aggregate result as reasoning ability. But many benchmarks are porous. They rely on familiar formats, public datasets, and narrow distributions. Models can look broad while still depending heavily on learned regularities.

There is also a deeper cognitive trap. Humans see reasons in narratives. We are storytelling creatures with a weakness for coherent sequences. When a model lays out steps that resemble our own classroom habits, we supply the missing mental furniture. We imagine the machine has an inner workspace because the transcript resembles what a competent person might write after using one.

None of this means the field learned nothing. Quite the opposite. Chain-of-thought prompting did improve task performance. Tool use, planning scaffolds, and verifier loops often help. The mistake was turning a practical technique into an ontological conclusion.

The product implications are larger than they look

If model reasoning is brittle in this way, several industry habits start to look shaky.

The first is the habit of treating polished explanation as evidence of reliability. In medicine, finance, law, security, or any workflow with real stakes, a fluent derivation should not buy much trust by itself. It might still be valuable as a draft, a hypothesis generator, or a way to expose assumptions for human review. But it should not be confused with stable competence.

The second is the way teams evaluate progress. If you only test on familiar benchmark distributions, you are mostly measuring how well the system navigates known grooves. That can still matter commercially. Lots of software succeeds by handling routine cases well. But you should describe that honestly. A dependable autopilot inside a narrow corridor is useful. It is not the same as a pilot.

The third is how companies think about customization. Fine-tuning can improve results dramatically for a particular workflow. There is nothing wrong with that. The danger is telling yourself a patched model has become broadly intelligent in the relevant domain. Often it has just become more specifically prepared.

This is where the human consequence shows up. People are already delegating judgment to systems that sound more certain than they are. A tool that mimics explanation can lower the user’s guard at exactly the wrong moment. The interface whispers, “You can stop checking now.” That whisper is one of the product features, whether anyone admits it or not.

Intelligence is still more than polished sequence prediction

The strongest version of the anti-hype claim would say these models never reason at all. The evidence does not force that conclusion. Some internal mechanisms may support limited abstraction. Larger models can display behaviors that look more compositional than smaller ones. External tools can also change the picture by offloading parts of the problem into explicit computation.

But the study does puncture a very popular shortcut: if the model can narrate steps, it must be thinking in rules. That shortcut no longer looks serious.

A better default is to treat these systems as powerful pattern engines with uneven pockets of generalization. They can be astonishingly useful inside those pockets. They can also wander outside them without noticing. The scary part is not that they fail. Everything fails. The scary part is how often failure arrives wearing the costume of understanding.

That should change how we use them. Ask for evidence, not just explanation. Perturb the wording. Test unfamiliar cases. Assume surface coherence proves less than it feels like it proves. If a model passes those harder tests, trust can rise on something sturdier than vibes.

The real shift is psychological. We do not need to be less impressed by what these systems can do. We need to be more precise about what, exactly, is doing the work when they do it.

End of entry.

Published April 2026