12 min read

The Closed-System Paradox: What Human Evolution Gets Right About the Future of LLMs

The fear is easy to summarize: once AI starts filling the internet with synthetic text, future models will train on model output, then on output from that output, and eventually drift into a mush of confident mediocrity. A photocopy of a photocopy. Detail fades. Errors harden. The weird edges of reality get sanded off.

That fear is real, but it also hides a stranger question. Human culture has always been recursive. Minds learn from minds. Books answer books. Most of what any one person knows arrived through language built by other people who learned through language built by others. No external superintelligence has been dropping fresh concepts into the species like software updates. And still, somehow, this closed conversational loop produced calculus, constitutional law, jazz harmony, orbital mechanics, and the recipe for puff pastry.

So why assume recursion is fatal for language models?

The answer sits in a distinction people blur too quickly. Human knowledge is recursive, but it is not merely self-referential. It is recursive under pressure from reality. That difference matters more than the purity of the data source.

The fantasy of pristine human data

A lot of discussion around model collapse quietly depends on a romantic idea: somewhere out there exists authentic human data, clean and original, and synthetic data is its degraded imitation. Once the fake overwhelms the real, intelligence decays.

That picture flatters us more than it explains anything. Human writing has never been pristine. It is stitched from imitation, memory, convention, misreading, and recombination. Plato was already packaging conversations into a philosophical form. Medieval scholars copied, glossed, and reshaped older texts. Novelists recycle plots that are older than nations. Most of language is inherited structure with local edits. The phrase “human-generated” sounds fresh only if you ignore how profoundly derivative humans are.

This is not an insult. Derivation is how culture works. Bach did not compose in a vacuum. Einstein did not wake up with relativity in a cave. Every serious intellectual achievement is both novel and saturated with precedent. Creativity is rarely ex nihilo. It looks more like compression plus mutation under constraint.

In that sense, synthetic text is not alien to the history of thought. It is another layer in a very old stack of minds remixing prior minds. If the critique is simply “these outputs are based on previous outputs,” humans are guilty on every count.

Yet the analogy only gets you halfway. The missing half is where the interesting part begins.

What model collapse really means

Model collapse is often described too vaguely, which makes it sound mystical. The core issue is statistical, not metaphysical. When a model is trained repeatedly on its own generations, it tends to overweight high-probability patterns and underrepresent rare but important features. Small distortions get fed back into the training loop. Over several rounds, the distribution narrows. You keep the average shape and lose the tails.

Imagine training a natural history illustrator using increasingly stylized drawings instead of real specimens. At first, the drawings look fine. The bird still has wings and a beak. By the fifth generation, every sparrow starts looking suspiciously like the average of all birds. Distinctive markings disappear. Anatomical mistakes become canon. The training set still contains “birdness,” but less contact with actual birds.

That is a real concern for language models because they learn from distributions, not from a direct understanding of why the text is true. If the data gets cleaner in the wrong way, the model may become more fluent and less grounded. It can grow more standardized while losing exactly the weird specificity that makes knowledge useful.

The important point is that recursive training is not always destructive. Synthetic data already helps in many settings. Self-play transformed game systems. Verified synthetic examples help in math and coding. Distillation can transfer useful behaviors from larger models to smaller ones. The difference is not whether data is synthetic. The difference is whether the loop contains corrective signals strong enough to stop error from becoming tradition.

Human culture is closed in one sense and open in another

This is where the human comparison becomes sharp. Human civilization is closed if you mean “we only get intelligence from other humans.” There is no outside teacher handing our species conceptual downloads. Progress emerges from internal exchange: conversation, argument, imitation, teaching, writing, institutions, memory.

But civilization is radically open if you mean “we are exposed to fresh information.” Reality keeps interrupting us. Babies are not trained on text alone. They have bodies, senses, needs, pain, weather, gravity, social pressure, scarcity, disease, boredom, embarrassment, hunger, deadlines, and other people rolling their eyes when their explanation makes no sense. The world is constantly labeling our outputs, usually without mercy.

That distinction wrecks the simplistic version of the paradox. Humans did not become more intelligent by passing symbols around inside a sealed library. We passed symbols around while living inside a stubborn universe. Ships sank. Bridges collapsed. Crops failed. Experiments refused to replicate. Predictions missed the mark. Sometimes an idea survived because it was beautiful. Much more often, ideas survived because they worked well enough that people kept them.

Take physics. Newton did not receive a cosmic data dump, but he did inherit centuries of observation, instrumentation, mathematics, and practical problems in navigation and astronomy. Einstein built on that archive, yet his theory gained authority because starlight bends, Mercury’s orbit shifts, and clocks disagree in measurable ways. Human thought is recursive, but the recursion keeps hitting the wall of the real.

Language models mostly do not have that wall. They are astonishing archive readers. Archive readers are useful. They are not the same thing as situated knowers.

The bottleneck is not recursion but uncorrected sameness

If recursion alone caused decay, human culture should have flattened long ago. Instead, it keeps generating novelty, sometimes by accident, sometimes by conflict, often by cross-pollination between traditions that previously ignored each other. The larger risk for AI is not circularity in the abstract. It is homogenization.

If the bulk of future synthetic text is produced by a handful of model families, tuned toward similar safety rules, similar styles, similar benchmarks, and similar reward signals, then the training pool becomes strangely monocultural. You get variety at the sentence level and conformity at the structural level. Millions of outputs may differ lexically while sharing the same habits of emphasis, omission, and caution. That is not a rich ecosystem. It is a plantation with good autocomplete.

Human cultures avoided this partly because they were fragmented for most of history. Languages diverged. Religions split. Isolated communities developed local practices. Rival schools defended incompatible premises for centuries. Heresies survived in margins. Translation scrambled meanings in productive ways. Even mistakes helped because different people made different mistakes.

That diversity was not always pleasant. It often came with war, exclusion, and preventable suffering. Still, from an information perspective, fragmentation preserved option value. The species did not place all its epistemic bets on one generator. A claim rejected in one region might thrive in another. A technique ignored by elites could survive in craft traditions. Useful knowledge often hid in places that looked unsophisticated to the center.

By contrast, current AI development has a strong centralizing tendency. A small number of labs gather most of the compute, much of the best talent, and huge chunks of public text. Their models are then used to generate even more text, code, summaries, and educational material. If those outputs become future training data, the loop is not merely recursive. It is recursively standardized.

Selection pressure is where humans still have the edge

Human knowledge progresses because culture is not just generated. It is selected. Ideas compete for survival in laboratories, courts, markets, classrooms, workshops, and daily life. Most proposals vanish. Some become doctrine too early and later get overturned. The filtering process is messy, political, and deeply imperfect, but it exists.

That selection is tied to consequences. A bad medical theory kills people. A weak bridge design cracks. An elegant theorem fails under proof. A clumsy interface annoys users until they leave for another tool. Reality keeps scoring our homework.

This matters because language-model training usually treats text as evidence of patterns, not as a history of consequences. A sentence in a forum post and a sentence in a lab notebook may both enter the corpus as token sequences. The model can learn stylistic and statistical differences, but it does not automatically inherit the world-level selection pressures that separated one from the other. Without additional machinery, all text starts to resemble testimony without a courtroom.

That is why the most promising path is not a desperate search for untouched “human” corpora. It is the creation of training loops that recover selection pressure. Some of that can come from tool use. Code either runs or it does not. A theorem checker accepts a proof or rejects it. A robot reaches the shelf or bumps into the shelf. Scientific workflows can compare predictions with observations. Software agents can be evaluated against objective tasks instead of applause.

Once synthetic data is tied to verifiable outcomes, it stops behaving like a hall of mirrors and starts looking more like experiment. AlphaGo’s self-play worked because the game supplies hard feedback. Programming agents improve when compilers and tests act as referees. Even language tasks become less slippery when the model has to retrieve evidence, act in an environment, or satisfy independent judges.

The human lesson is not “self-reference is safe.” The lesson is that self-reference becomes productive when coupled with selection.

Time and scale change the picture

There is another reason the human analogy can mislead if used carelessly. We have had enormous amounts of time. Cultural evolution ran for millennia under broad diversity and endless local feedback. Many bad ideas persisted for centuries before breaking. Many good ones took generations to spread. Human progress looks smooth only from far away. Up close, it is a chaos machine with libraries.

Language models operate on compressed timelines. A model can absorb a civilization-scale corpus in weeks, then shape global behavior within months. That speed is a gift and a hazard. It means error can scale before selection has time to correct it. If millions of people rely on AI-generated explanations, tutorials, and summaries, the model is no longer just learning culture. It is actively editing the future training distribution in real time.

This feedback loop could get ugly in obvious ways. Search results might fill with polished but derivative articles. Educational material might converge on the same simplifications. Code repositories could accumulate statistically normal but architecturally mediocre patterns. The issue would not be dramatic collapse into nonsense. It could be subtler: widespread competence paired with thinning originality.

Humans have their own version of this problem. Mass schooling, standardized testing, corporate communication, and platform incentives already flatten expression. AI may accelerate a tendency that predates it. That matters because the benchmark for healthy knowledge systems is not some mythical era of untouched human originality. It is whether a culture preserves enough variation and enough reality contact to keep discovering things.

Building better loops

The practical implication is surprisingly clear. The future of advanced language models will depend less on preserving a sacred reservoir of purely human text and more on designing ecosystems where generated knowledge can be challenged, split, recombined, and tested.

That means more pluralism in model development, not less. Smaller specialized models, regionally tuned models, scientific models, coding models, models trained on different objectives, and models exposed to different environments all reduce the risk of monoculture. It means keeping provenance so training pipelines can distinguish verified synthetic examples from generic fluent sludge. It means linking generation to external checks wherever possible. It also means remembering that some of the most valuable future data will not look like prose at all. It will be traces of action: experiments, simulations, transactions, tool calls, edits, failures, and outcomes.

There is an institutional angle too. Human knowledge improved through systems that preserved disagreement while rewarding contact with evidence. Peer review is flawed, markets are noisy, and democratic deliberation is often exasperating, but they all create friction. Friction is useful. A world where AI agents endlessly agree with one another in the same smooth dialect may feel efficient and produce very pretty dashboards. It will not be epistemically healthy.

The strongest models of the next decade may look less like giant frozen readers of the web and more like participants in ongoing environments. They will read, generate, test, revise, call tools, compare hypotheses, and learn from the consequences of action. In that world, synthetic data is not the enemy. Synthetic data without fresh constraint is the problem.

The closed-system paradox resolves in the open world

The interesting part of the original paradox is that it starts from a correct intuition. Intelligence does not require a magical source of pure novelty. Human beings prove that recursive exchange can generate extraordinary new structure. Most ideas are born from previous ideas. In that narrow sense, the panic over “AI trained on AI” is too crude.

But the same human example also shows why the crude optimism fails. We did not progress because recursion alone is enough. We progressed because recursive culture stayed entangled with bodies, tools, environments, rival communities, and consequences. Our loop was social, but it was never sealed.

That should shift the question. The problem is not how to protect an untouched reservoir of human text from contamination by machine output. That reservoir never really existed. The problem is how to build systems that keep encountering novelty they did not author and standards they cannot negotiate away. A model that reads only generated language is like a scholar locked in a library with no windows. A model that acts, checks, fails, and tries again starts to participate in the same broader pattern that made human culture cumulative.

If future language models stagnate, it will not be because synthetic data is inherently corrupt. It will be because we built a narrow loop, centralized it, polished it, and mistook fluency for contact with the world.

End of entry.

Published April 2026