Calibrated Failure: What CASP13 Teaches Us About AI Innovation
DeepMind won one of biology’s most important competitions in 2018 and left feeling deflated. That reaction matters more than the trophy.
CASP13, the biennial protein-structure prediction contest, was supposed to validate years of work. DeepMind’s AlphaFold system beat the next-best team by a wide margin, roughly 50% in the hardest category. From a distance, it looked like a clean story about machine learning arriving in science and taking over.
Up close, it was the opposite. The team had proof they were ahead of everyone else, and almost no proof they had solved anything a working biologist could use. John Jumper later put it in a line that should be engraved above every benchmark dashboard in Silicon Valley: “We were the best in the world at a problem the world’s not good at.”
That sentence captures a hard fact about AI progress. Winning is not the same as being useful. A large lead over weak baselines can still leave you nowhere near the threshold that reality demands.
A win that landed like bad news
CASP exists because protein folding is both scientifically central and technically punishing. Given a sequence of amino acids, can you predict the 3D structure a protein will adopt? Do that reliably and you can accelerate everything from basic biology to drug discovery. Fail, and you are back to expensive experiments, slow iteration, and years of uncertainty.
The competition is not a toy exercise. Organizers collect protein structures that have been solved experimentally but not yet made public. Teams submit predictions blind. Later, judges compare those predictions to the real structures. It is one of the cleaner interfaces between machine learning and a consequential scientific task. There is a leaderboard, yes, but there is also a real-world standard hiding underneath it.
AlphaFold 1 did very well by contest standards. DeepMind used deep learning to infer distances between amino acids and combine those estimates with physical and geometric constraints. It was clever, ambitious work. It improved the state of the art.
Then the biologists weighed in.
Janet Thornton, a major figure in structural biology, said the quality of the predictions varied and they were “no more useful than the previous methods.” Nobel laureate Paul Nurse was even more direct: AlphaFold did not produce data good enough to be useful in practical research.
That contrast is the whole story. Inside the machine-learning frame, DeepMind had won. Inside the biology frame, the output still did not clear the bar. If you only look at relative improvement, you miss the more important variable: the size of the remaining gap.
This is common in AI. A model can halve error rates, set records, and still force the user to do the same old manual workaround. The graph looks dramatic. The lived experience barely moves.
Relative progress can hide absolute failure
Benchmarks train teams to think in deltas. Did we beat the prior system? Did we widen the gap with competitors? Did our score move enough to justify the cost? Those are reasonable questions. They become dangerous when they replace the real one: is the output good enough to change what someone can do?
Protein prediction makes this painfully visible because biology has little patience for “almost.” If a predicted structure is off in a way that distorts the active site or misplaces a key interaction, a researcher can waste months following the wrong lead. You do not get partial credit in the wet lab because the model was better than last year’s model.
That is why CASP13 matters beyond biology. It exposes a standard failure mode in AI organizations: substituting comparative excellence for task completion. The team can be elite, the engineering can be beautiful, the benchmark win can be real, and the product can still be unusable.
Jumper’s other image is even sharper: “It doesn’t help if you have the tallest ladder when you’re going to the moon.” You can feel the frustration inside that metaphor. The issue is not effort. The issue is category error. You are optimizing an approach that cannot reach the destination.
Richard Evans, who worked on the project, later said the team had thought they could throw some of their best algorithms at the problem. “We were slightly naive.” That kind of naivety is everywhere in AI. Strong general methods tempt teams into thinking every stubborn domain is mostly waiting for enough compute, enough data, or enough persistence. Sometimes that is true. Often it is only partly true. And the expensive part is discovering which kind of wall you are hitting.
There are at least two kinds of resistance in technical work. One comes from scale. The method is basically sound, but the system needs more data, more training, more engineering discipline, or better hardware. The other comes from mismatch. The method is pointed at the problem from the wrong angle, so scaling it mostly produces a cleaner version of the same insufficiency.
CASP13 looks, in retrospect, like a mismatch disguised as progress.
Biology supplied the only metric that mattered
The most impressive thing about the AlphaFold story is not the final breakthrough. It is that DeepMind listened when domain experts said the win did not count yet.
That sounds obvious. It is not. Modern AI culture has a bad habit of assuming the benchmark is the truth and the user is confused. When that happens, criticism gets translated into “adoption lag,” “change management,” or some other polite phrase that means reality has not updated itself to the demo.
Structural biology did not offer that comfort. Either the predictions were reliable enough to support scientific inference, or they were not. Thornton and Nurse were not saying the work lacked promise. They were saying promise was not the unit they needed.
This distinction matters because frontier AI now lives in a fog of proxy metrics. Models ace standardized tests, produce eloquent text, and solve polished coding tasks. Then they enter a company, a hospital, a lab, or a legal workflow and suddenly need constant supervision. The benchmark has measured competence in a narrow theater. The user is measuring whether the system can carry weight without hidden damage.
That is why the AlphaFold team’s discouragement was healthy. They treated criticism as instrumentation, not as an insult. Many organizations never make that transition. A flattering metric becomes part of the company’s identity, and then defending the metric matters more than interrogating its relation to the task.
In science-facing AI, that attitude collapses quickly because the world gives hard feedback. A structure is either near experimental quality or it is not. In consumer software, the feedback can be slower and easier to spin. A support chatbot that resolves 70% of tickets sounds efficient until you learn the remaining 30% contain the legal threats, refund escalations, and account lockouts. A coding agent that passes benchmark suites sounds magical until your senior engineers spend afternoons cleaning up subtle errors with all the charm of a Roomba dragging mud across the living room.
The user is often holding the only honest metric in the room.
Timing decides whether ambition survives
Demis Hassabis drew a wider lesson from the AlphaFold journey: ambition is good, but timing matters. There is no point being fifty years ahead of your time, he said, because you will never survive fifty years of that effort before it pays off.
That comment lands because it is not anti-ambition. It is about survivability. A grand challenge is not just a scientific target. It is a bet that enough adjacent pieces are ready, or about to be ready, for your work to compound rather than evaporate.
This is one of the least romantic truths in research. Timing is not a side detail. It changes what counts as vision and what counts as martyrdom.
Protein folding had resisted decades of work because several ingredients were missing or immature at once. There were not enough powerful learning architectures. Compute was weaker. Biological data pipelines were rougher. Integrating machine learning with structural biology still required crossing a cultural border, not just a technical one. By the late 2010s, enough of those constraints had shifted that a real attack became plausible. A few years earlier, the same level of ambition might have produced mostly burn and little traction.
People often talk about timing in startup language, as if it means market readiness and investor mood. In frontier AI, timing is more physical. Are the representations expressive enough? Is the data substrate rich enough? Do you have a way to close the loop with domain experts? Can the organization afford a detour that does not immediately monetize? If several answers are no, brilliance alone does not rescue the effort.
That is why “just keep going” is weak advice in these contexts. Persistence matters, but persistence without calibration can become a slow loyalty to the wrong method. The discipline is to tell whether the field is immature, the architecture is wrong, or your expectations are mis-set. Those diagnoses look similar in the middle of the struggle. They only look obvious in documentaries.
The rebuild after CASP13
DeepMind’s response to CASP13 was not to squeeze a few more points out of the same recipe. They rebuilt.
That decision deserves more attention than the celebrated CASP14 result. Rebuilding after failure is common. Rebuilding after a public win is rare. It requires a group to say, in effect, yes, we outperformed everyone, and no, that performance does not justify attachment to our current stack.
The changes were substantial. The team rewrote the data pipeline. They brought biologists deeper into the work, not as decorative advisors but as people shaping what the system needed to learn. Kathryn Tunyasuvunakool described the goal plainly: biological relevance. That phrase sounds modest. It was actually a demand to realign the whole project.
AlphaFold 2 moved away from a more piecemeal pipeline toward an end-to-end architecture that could reason over relationships within protein sequences and across evolutionary data more directly. Its internal machinery, especially the Evoformer component, used attention-like operations to update sequence and pair representations together. The important point is not the nomenclature. The important point is that DeepMind stopped treating the problem as something to be solved by bolting better machine learning onto an inherited decomposition. They changed the decomposition.
That shift is easy to miss when people retell the story as “then transformers showed up.” Architecture mattered, but so did the decision about what the model should predict, how the data should be represented, and how tightly domain knowledge should be woven into the system. There is no pure scaling story here. There is a story about choosing a representation that let the model see the structure of the problem more faithfully.
CASP14 made the difference unmistakable. AlphaFold 2 achieved accuracy levels many researchers had barely allowed themselves to expect. John Moult, the organizer of CASP, wrote to the team that their performance was amazing both relative to other groups and in absolute terms. That second clause is the one that matters. Absolute terms. The model had crossed into usefulness.
Thornton, who had criticized earlier results, later said this was a problem she had begun to think would not be solved in her lifetime. The same domain that had dismissed the earlier outputs now recognized a genuine step change. That is what it looks like when a benchmark catches up with the task.
Calibrated failure is a real capability
The phrase I keep coming back to is calibrated failure. Some failures only produce noise. Others tell you exactly how far you are from the bar and whether your current direction deserves another year.
CASP13 was valuable because it was painful and legible at the same time. DeepMind did not merely learn that AlphaFold 1 was imperfect. They learned that relative superiority had masked absolute inadequacy. That gave them a map, or at least a boundary.
The best AI teams increasingly need this capability because the easy era of benchmark worship is ending. Many important tasks now have thresholds that are steep rather than smooth. A translation system that gets the gist is useful for reading menus and terrible for contracts. A coding model that handles routine functions is helpful until it starts modifying authentication logic. A medical summarizer that usually captures the chart is one hidden omission away from a dangerous hallucination. In these settings, the path from 70 to 90 is not a 28% improvement. It is the difference between “interesting” and “deployable.”
That changes how organizations should interpret progress. If you are far below a threshold, a leaderboard gain can still be strategically meaningless. If you are close to a threshold, a smaller gain can unlock a flood of real use. The shape of the utility curve matters more than the headline score.
It also changes how you staff teams. The AlphaFold story is often framed as machine learning defeating a classic scientific problem. A more accurate reading is that machine learning became effective once it was forced into a productive relationship with the domain. When Tunyasuvunakool says biological relevance drove the pipeline rewrite, that is not an implementation footnote. It is the central move. The experts closest to the consequences of error helped redefine what the model was for.
This is where many AI programs still wobble. They hire domain specialists late, ask them to validate demos, and then seem surprised when the demos do not survive contact with practice. The useful arrangement is earlier and messier. The model builders need people who understand where failure actually hurts, which parts of the output matter disproportionately, and what “good enough” means once the slide deck disappears.
Calibrated failure also demands emotional maturity. A team has to absorb the possibility that its most celebrated result is mostly diagnostic. That is hard for obvious reasons. Careers, funding narratives, and public expectations all prefer cleaner arcs. Yet the alternative is worse. You can spend years polishing a system that is admired in conference halls and quietly avoided by the people it was built for.
Benchmarks are maps, not territory
None of this means benchmarks are bad. CASP itself is proof that careful evaluation can move a field forward. The problem starts when the score becomes more real than the task.
A good benchmark is a map. It compresses a complicated territory into something teams can compare, optimize, and discuss. That is useful because the world is noisy and progress needs common reference points. But a map is selective by design. It preserves some features and discards others. If you forget that, optimization becomes a kind of cartographic superstition.
The current AI industry is full of teams standing very proudly on maps.
Reasoning benchmarks reward polished solution traces while telling you little about whether a model can manage a week-long research project. Coding benchmarks capture correctness on bounded tasks while missing maintenance burden, integration risk, and the strange human labor of reviewing machine-written code that is almost right. Voice assistants can sound uncannily natural while still failing the social contract of a phone call, where confidence without reliability is more irritating than obvious incompetence.
CASP13 offers a better posture. Celebrate progress if it is real, but force every evaluation to answer two questions. How much better are we than alternatives, and how far are we from the level that changes the user’s world? Those questions can point in different directions. When they do, the second one should have veto power.
That is not anti-science, anti-benchmark, or anti-ambition. It is a way of preserving ambition from vanity. The point of aiming at a grand challenge is not to dominate the intermediate scoreboard forever. It is to know when the scoreboard is helping and when it has started to flatter you.
The value of a painful benchmark
The reason the AlphaFold story still hits is that its pivotal moment was not the triumph everyone remembers. It was the disappointment that came first.
DeepMind’s CASP13 result was a kind of gift wrapped as embarrassment. It revealed that the team had enough skill to lead the field and still not enough to matter to biology. Many groups would have treated that as a branding problem. They treated it as a measurement problem, then an architectural problem, then an organizational problem. That sequence is why CASP14 became possible.
There is a broader lesson here for AI innovation. Progress is not just about moving faster than peers. It is about staying honest about the distance left to usefulness, especially when your current metrics are cheering. A benchmark win can be evidence of momentum. It can also be evidence that your entire field has not yet found the right handle on the task.
The difference is not visible in the confetti. It becomes visible when the people closest to the real work say, with irritating precision, that your impressive result still does not help them enough. The teams worth paying attention to are the ones that can hear that sentence early enough to start over.
End of entry.
Published April 2026