13 min read

The Paradox of Fidelity: Why AI Needs Errors to Innovate

The campaign against weirdness

Every major AI lab is trying to make its models less strange.

That effort makes sense when the model is drafting a legal memo, summarizing a lab result, or helping someone change a prescription. In those settings, hallucination is not charming. It is a liability with a user interface. The industry has spent two years turning probabilistic text generators into systems that behave more like careful interns and less like sleep-deprived improv comics.

But there is a cost hiding inside that cleanup operation. If you squeeze out too much error, you may also squeeze out a major source of novelty.

Joseph Henrich, who studies cultural evolution, makes the point more bluntly than most AI people do. Across human history, a surprising number of innovations came from copying mistakes. Somebody reproduced a technique imperfectly and got a better result. A tool balanced differently. A recipe fermented longer. A ritual changed shape and ended up coordinating a group more effectively. Evolution, cultural or biological, does not move by perfect transmission alone. It moves because fidelity and variation sit in tension with each other.

That tension matters for AI because we are building systems that excel at high-fidelity reproduction. They absorb patterns at huge scale, compress them, and replay them with astonishing smoothness. The question is whether smoothness is always what you want.

Innovation is often miscopying with a selection layer

People like to talk about invention as if it arrives fully formed in a genius brain. History is less cinematic. Much of it looks like apprenticeship, imitation, drift, and accident.

A craftsperson learns from another craftsperson. A builder copies a joint imperfectly. A cook substitutes one ingredient because the usual one is unavailable. Most of those deviations fail. Some are useless. A few turn out to be better adapted to local conditions than the original. Then others copy the improved variant, imperfectly again, and the cycle continues. Cultural evolution is not a parade of pristine ideas. It is a filtering system for deviations.

Biology gives us a cleaner version of the same logic. DNA replication is extraordinarily accurate, but not perfectly accurate. If it were perfect, adaptation would stall. If it were wildly inaccurate, useful structure would dissolve. Life survives in the narrow band where information is preserved well enough to persist and altered enough to explore. Some parts of genomes are more mutable than others because evolution itself has, over long periods, tuned where variation is more tolerable.

The analogy to AI is not exact, but it is revealing. We are used to treating model errors as pure failure because we evaluate them in settings where correctness is binary and immediate. Did it cite a real case. Did it return the right dosage. Did it identify the company revenue correctly. In those tasks, the answer really is yes or no, and the “creative” error is just an error.

Yet innovation rarely begins in that mode. It begins upstream, in an exploratory phase where the system is generating options that will later be filtered by experiment, criticism, or reality. Human societies have always relied on this asymmetry. We tolerate a lot of bad variation because selection happens after generation, not before it.

The modern AI stack is unusually focused on the first half of that equation. We generate with power, then spend enormous effort suppressing deviations before they appear. That can produce cleaner answers. It can also produce a machine that keeps reaching for the center of its distribution, which is a polite way of saying it gets very good at sounding like what already worked.

The average answer is often the safe answer

A large language model is trained to predict likely continuations. Fine-tuning then pushes its behavior toward what humans rate as helpful, harmless, and reliable. Retrieval systems add factual scaffolding. Safety filters shave off more variance. Product teams run evaluations that reward consistency, low surprise, and reduced liability.

This stack is rational. It is also deeply conservative.

The safest answer is usually the answer closest to the existing record. When you ask for a strategy memo, a product concept, or a scientific hypothesis, the model has structural incentives to give you something that resembles what the training distribution already considers plausible. Even with a high temperature setting, most of the weirdness it produces is not the kind you would call generative in a useful sense. It is often decorative novelty rather than conceptual novelty.

Humans are not magically better. We also copy, conform, and repeat ourselves. The difference is that human cultures are full of broken transmission channels. People mishear. Teams work from incomplete context. Disciplines remain partially isolated. Two firms share a cafe but not a roadmap. A junior engineer carries an idea from one domain into another without realizing it is “out of place.” Those frictions are inefficient in many ways, yet they create room for combinations that no centralized system would deliberately plan.

A model that can “see everything” may actually lose some of that productive mess. Total visibility tends to collapse differences into statistical regularities. If every interesting pattern is instantly legible and recombined into the same giant predictive surface, novelty risks becoming a style transfer problem. You get endless variation in expression while deeper assumptions remain fixed.

This is one reason AI-generated output often feels broad but oddly flattening. It can synthesize across domains, but synthesis is not the same as surprise. Many outputs feel like composites of what a smart generalist would already think. Useful, sometimes excellent, but rarely catalytic.

Hallucination is not creativity, but it can feed it

There is a tempting line people use in this debate: hallucination and creativity are basically the same thing. That is too loose to be trusted.

When a model invents a source in a court filing, that is not creativity. It is system failure. When it fabricates a biography, users do not need a more poetic definition. They need the model to stop. Calling every false statement “creative variation” confuses exploratory generation with tasks where the world pushes back immediately.

Still, the instinct behind the line is not entirely wrong. In creative work, scientific ideation, and early-stage design, a model’s willingness to produce things that do not yet exist can be genuinely useful. The problem is not that the system departs from the record. The problem is that we often ask it to depart and verify at the same time.

Those are different jobs.

A chemist brainstorming candidate compounds can use a model that wanders, provided later stages test viability. A game designer may benefit from mechanics that are improbable or structurally odd. A writer can use a model that makes lateral jumps between metaphors, references, and forms. In each case, the value comes from separating generation from adjudication. First you let the machine produce candidates. Then another process, human or computational, checks which candidates survive contact with evidence, constraints, or taste.

Humans do this constantly. The first idea in a notebook is not the final argument in a paper. A sketch is not a blueprint. A rumor is not a measurement. Our institutions for discovery exist partly to turn noisy generation into dependable knowledge. Labs, editors, prototype cycles, and peer review all perform selection. Without that second layer, variation is just clutter.

AI products often collapse these stages into one interface. The same chat box is expected to brainstorm, explain, fact-check, and execute. Then we act surprised when its personality becomes contradictory. A system optimized for maximal reliability will underperform at imaginative search. A system optimized for adventurous search will occasionally hand you nonsense. Treating those goals as identical is a category mistake disguised as product design.

Serendipity depends on partial overlap

Henrich points to research from Silicon Valley showing that firms which frequent the same coffee shops are more likely to cross-reference each other’s patents. That finding sounds almost quaint next to trillion-parameter models and hyperscale compute, but it captures something the software worldview often misses.

Innovation is not just the recombination of all available knowledge. It is the recombination of knowledge under conditions of limited access, local context, and accidental encounter.

The coffee shop matters because it is not a meeting with an agenda. It is a place where overlap happens without full coordination. People hear fragments. They misinterpret each other slightly. They connect ideas that a formal collaboration would keep in separate folders. The scene is messy, but the mess is doing work. It creates collisions between partially shared worlds.

AI can, in principle, meet everyone at once. That sounds like an advantage. Why rely on coffee shops when a model can ingest papers, patents, source code, and internal documentation from thousands of domains in one training run?

Because total access changes the shape of discovery.

If every system draws from the same corpora, is tuned on the same benchmarks, and is steered by similar safety and product incentives, it may produce global coherence at the expense of local eccentricity. You get broad interoperability, which is valuable. You also get convergence. The same assumptions harden across industries because the same model families mediate the first pass of thought everywhere. A coffee shop exposes you to adjacent minds. A foundation model can expose everyone to the same synthesized median.

There is another wrinkle. Serendipity is not randomness alone. It is randomness filtered by proximity. The cafe down the street produces different collisions than a global firehose because the people in it share enough context for an accidental idea to become usable. Total recombination can be too diffuse. You can drown in possible combinations that no one in a specific setting can actually apply.

That suggests a counterintuitive design principle. The goal may not be to maximize access to every idea at once. It may be to preserve local variation and incomplete overlap so that different groups explore different parts of the landscape. AI that homogenizes discovery could make everyone smarter in the same direction, which is a narrower gain than it first appears.

Stable systems adapt poorly to shocks

Henrich worries about systems that become too homogeneous and lack enough noise to generate new challenges. That concern lands harder when you stop thinking only about creativity and start thinking about resilience.

Human societies learn under pressure. Recessions change behavior. Pandemics reveal dependencies. Climate events expose weak infrastructure. Wars, migrations, and supply shocks all inject new information into the social system whether we welcome it or not. These shocks are costly and often tragic, but they force adaptation by making old assumptions visibly wrong.

Machine learning systems are built to reduce variance. That is their superpower. It is also a hazard in dynamic environments.

A model trained on years of stable historical data may perform beautifully until the world shifts. A recommendation engine tuned for normal consumer behavior can misread a sudden supply disruption. A coding assistant can reinforce best practices from yesterday’s stack while missing an architectural break underway. A planning model can optimize for the benchmark it knows while real constraints move under its feet. If many institutions depend on similar systems, they can fail in correlated ways.

This is not a hypothetical edge case. Finance has lived with versions of this problem for decades. When many actors optimize against similar signals, they can all become efficient in the same direction and fragile in the same moment. AI raises the possibility of that pattern spreading far beyond markets. If the same model families help write policy drafts, evaluate resumes, generate software, summarize research, and shape strategic decisions, then a hidden bias or blind spot stops being local. It becomes infrastructure.

Variation helps because it prevents synchronized overconfidence. Different models, different data windows, different fine-tuning choices, and different institutional contexts create disagreement. Disagreement is inconvenient. It is also a way of sensing the edges of what you do not understand.

Productive error has to be designed, not romanticized

None of this means we should celebrate every model mistake as a seed of genius. Most mistakes stay mistakes. Anyone who has spent time with language models knows how much low-value noise they can generate when left unchecked. The romantic story, where every hallucination hides a breakthrough, is just the inverse of the compliance story where every hallucination is a defect. Both flatten the issue.

The useful question is narrower: where do you want exploration, and where do you require precision?

In medicine, dosage calculation should be boring. In legal citations, boring is beautiful. In accounting, the ideal model has the personality of a calculator with good manners. But upstream from those tasks, in hypothesis generation, scenario planning, interface design, or speculative research mapping, a certain amount of structured deviation is desirable. You want a system that can leave the groove, provided the next stage catches nonsense before it reaches consequence.

That implies a different architecture for AI work. Instead of demanding one immaculate model that does everything, organizations may need paired systems and paired norms. One layer explores broadly, with permission to generate unlikely candidates. Another verifies, narrows, and grounds. One model can even play both roles in sequence, but the distinction between them has to be explicit. Otherwise, the same tool will be judged as both too cautious and too unreliable, often in the same afternoon.

There are technical ways to support this. Ensemble methods, diversified prompting, model disagreement analysis, synthetic counterfactuals, and domain-specific simulation can all create bounded variation. Yet the institutional layer matters more than the trick. Teams need workflows that know when a weird answer is material for testing and when it is simply disallowed. That boundary is less glamorous than the model itself, which is why product marketing tends to skip it. The boundary is where trust actually lives.

The real challenge is selection

Human innovation did not emerge because people were gloriously error-prone. It emerged because societies developed ways to keep useful deviations and discard the rest.

That is the part AI discourse often underrates. Generation is abundant now. Selection is scarce. We can already ask a model for fifty business ideas, twenty mechanism designs, ten molecules, or a hundred variants of a user flow. The bottleneck is no longer producing possibilities. It is evaluating them cheaply, fairly, and in context. A hallucination becomes valuable only when some downstream process can detect whether it points toward a viable path.

This shifts the center of gravity from model purity to system design. The important question is not whether AI can make mistakes. It obviously can. The question is whether our institutions know how to turn some of those mistakes into experiments instead of liabilities. Science can sometimes do that. Design can often do it. Bureaucracy usually cannot. Regulated industries need careful partitioning. Consumer products need guardrails that match the stakes. The answer will vary by domain, which is inconvenient for anyone hoping one alignment recipe will solve the whole field.

A perfectly faithful machine sounds like the end state of progress because fidelity is easy to score. It is measurable, legible, and marketable. But history suggests a stranger lesson. Systems that only preserve what already works tend to become elegant archivists of a world that is already changing. If AI is going to help produce genuinely new knowledge, it will need room to generate departures from the record, and we will need better ways to decide which departures deserve a second look.

End of entry.

Published April 2026