From Scaling to Research: AI’s 2025 Turn
“There are more companies than ideas.”
Ilya Sutskever’s line lands because it reverses the story the industry has been telling itself. For five years, the winning move looked almost insultingly simple: buy more chips, gather more data, train a bigger model, repeat. If you had enough money and enough patience, progress felt scheduled.
That story is ending, or at least losing its monopoly on the future.
Sutskever’s claim is not that compute stopped mattering. Quite the opposite. He is talking about a return to fundamental research with gigantic machines in the loop. The difference is subtler, and more important. Between 2020 and 2025, scaling became a recipe. In the next phase, compute remains essential, but it stops telling you what to do. The scarce input shifts from hardware to insight.
That changes the mood of the entire field. It also changes who gets to win.
The five-year interval that looked like destiny
It is easy to forget how contingent the scaling era really was. Once GPT-3, then GPT-4-class systems, started showing broad capability gains from larger training runs, the industry snapped into a kind of industrial logic. Scale was no longer a hypothesis. It was a procurement plan.
This was catnip for companies. Research is moody. It advances through false starts, half-formed intuitions, ugly dead ends, and occasional miracles. Scaling, by contrast, looked legible to a board. Spend more, get more. The output might not be perfectly predictable, but it was predictable enough to finance. Datacenter construction became a strategy slide instead of a support function.
That is one reason the period from 2020 to 2025 felt so uniform. Labs that had different cultures, different origin stories, and different public rhetoric increasingly converged on the same behavior. Raise capital. Secure compute. Hoard data. Train. Ship. Train again. Even the disagreements happened inside the same frame.
Sutskever’s description of the period is sharp: scaling “sucked all the air out of the room.” Once that happens, imitation starts masquerading as consensus. Everyone does the same thing partly because it works, and partly because no one wants to explain to investors why they are doing something harder to measure.
The result was real progress. It would be silly to deny that. But the comfort of the recipe also hid a deeper truth. A recipe is powerful precisely because it lets you postpone the next conceptual leap.
Three phases, one industry
Sutskever divides recent AI history into three phases, and the frame is useful because it shows what changed.
From roughly 2012 to 2020, the field was driven by research breakthroughs. AlexNet demonstrated what deep learning could do for vision with what now looks like comically modest hardware: two GPUs. ResNet changed how very deep networks could be trained. The Transformer, introduced in 2017, reshaped the path of the whole field with a training setup tiny by current standards. These advances were not produced by brute force alone. They were conceptual moves.
From 2020 to 2025, scaling took over. The core insight of the era was that larger models trained on more data with more compute kept getting better in smooth, useful ways. The scaling laws were not merely an academic observation. They became an industrial doctrine. If loss curves keep rewarding expenditure, then capital has a direct route into capability.
The third phase, in Sutskever’s telling, starts around now. We are heading back toward a research-led era, except this time researchers operate in an environment filled with giant clusters. That matters. The return is not nostalgic. Nobody is going back to proving the future on a couple of consumer cards in a cluttered lab. The difference is that compute becomes a proving ground for ideas rather than the idea itself.
This framing helps explain a tension many people in AI have been feeling without naming clearly. The field still feels expensive, but less intellectually settled. The largest labs are spending more than ever while also sounding less certain about what the next leap will be. That is exactly what a transition between paradigms looks like.
Scaling solved management before it solved intelligence
The simplest way to understand the appeal of scaling is that it solved a management problem.
If you run a frontier lab or fund one, research uncertainty is painful. A brilliant team can work for eighteen months and produce nothing of strategic value. You can hire exceptional people and still miss the next important idea. That is normal science. It is much less comfortable as a business process.
Scaling softened that uncertainty. When Sutskever says companies liked scaling because “you know you’ll get something,” he is naming the real breakthrough of that period. Scaling translated an unruly research field into something that could be budgeted, forecast, and operationalized. You did not need to know exactly what new capability would emerge. You only needed confidence that enough additional compute would buy some meaningful improvement.
That is a very different promise from traditional research. It resembles building faster logistics infrastructure or a larger refinery. Hard, expensive, competitive, but intelligible.
Once a field starts behaving that way, the center of gravity shifts. The people with the most influence are no longer only the people with the best ideas. They are also the people who can assemble supply chains, financing structures, energy contracts, and political relationships. AI stops being merely a scientific race and becomes a construction project.
That did not make the science fake. It did make the dominant skill set broader and, in some corners, blunter. There is a reason the AI boom produced so many conversations about GPUs, fabs, power availability, and sovereign compute. Those were not side issues. They were the terrain.
The limits are showing in public
The scaling recipe is not collapsing in a dramatic Hollywood explosion. It is hitting limits that look more mundane and therefore more consequential.
One is data. Sutskever’s point is direct: pretraining data is finite. The internet is not an infinite well of high-quality text, images, code, and multimodal interactions. You can improve data pipelines, curate better corpora, synthesize new material, and squeeze more value from existing sources. But the old assumption that the next order-of-magnitude training run can simply ingest a much larger universe of useful information is getting shakier.
Another limit is the character of the gains themselves. More compute still changes models. It still buys performance. But Sutskever questions whether multiplying training compute by 100 would transform everything in a fundamental sense. His answer is basically no. The system would be different, likely better in many ways, perhaps materially better. Yet “different” is not the same as a qualitative shift in what the model can do, how robustly it can reason, or how safely it can operate.
That distinction matters because the industry has spent years blurring it. Bigger models did produce surprisingly discontinuous-feeling user experiences. A chatbot that can summarize a legal document, write working code, or tutor a student feels categorically unlike the toy models that came before. But under the hood, many of those jumps emerged from smooth curves crossing practical thresholds. Once those thresholds are crossed, it is tempting to assume the next leaps will work the same way forever.
They may not.
There is also an awkward empirical fact here. Some of the most interesting recent advances have involved changes in training regimes, reasoning scaffolds, synthetic data generation, tool use, and inference-time search. Those are not arguments against scaling. They are hints that raw pretraining scale is no longer enough to carry the narrative by itself.
The scarcest input is now conceptual
Sutskever puts the point almost provocatively: if ideas are so cheap, why doesn’t everyone have them?
That question cuts through a lot of startup mythology. In software, we like to say ideas are abundant and execution is rare. In frontier AI, that slogan suddenly looks less universal. Execution still matters enormously. But if dozens of well-funded labs with access to world-class talent and serious compute are all chasing the next breakthrough, then the shortage is not ambition. It is genuine novelty.
This is where the historical examples matter. AlexNet did not need the largest computing cluster on Earth. The Transformer did not require a nation-state budget to prove itself. Even more recent reasoning-oriented work, including ideas associated with systems like o1, was not born from simply turning one familiar dial harder than everyone else. In each case, the important step was conceptual. Once seen, it looks obvious in hindsight, which is one of research’s recurring jokes.
The phrase “ideas are the bottleneck” can sound romantic, as if the field is returning to pure thought after an unfortunate flirtation with infrastructure. That would be misleading. Big ideas in AI are expensive to test now. The point is not that compute vanishes. It is that compute has become less discriminating. Many frontier players can access enough of it to validate a serious idea. The real scarcity lies upstream, in forming the idea worth validating.
That shifts the internal economics of a lab. If a hundred million dollars can no longer guarantee a strategically decisive leap, then taste, patience, and scientific leadership matter more again. Some organizations are built for that. Others were built to capitalize on a trend line.
What a research-era company actually looks like
Sutskever’s own company, Safe Superintelligence Inc., is an interesting case because it is designed around this thesis. Its pitch is not that it can ship the slickest product this quarter, or monetize inference most efficiently next quarter. Its advantage, as he frames it, is structural concentration. It does not need to split compute across consumer features, enterprise support, model serving, and a dozen adjacent bets. It can direct resources toward foundational work.
That is not a small distinction. Every major AI lab that also runs a large product business lives with a tension between research purity and operational gravity. Inference demand is endless. Customers always want lower latency, higher reliability, more integrations, better pricing, and special fine-tunes for their workflow. Product success pulls compute toward service. Service, in turn, shapes research agendas. You start optimizing for what can be shipped and monetized, not necessarily for what could open the next frontier.
A focused research company can escape some of that drag. It can tolerate longer cycles. It can invest in ideas that have weak short-term narratives. It can also make stranger bets without immediately justifying them to a paying customer base.
Of course, there is a catch. A pure research structure gives you freedom, but it also removes the disciplining effects of real use. Product companies learn from user behavior at scale. They see failure modes in the wild. They discover where models are brittle, deceptive, useful, boring, or economically transformative. That feedback is not trivial. A lab insulated from products can end up elegant and detached, like a Formula 1 team that forgot roads exist.
The likely outcome is not that one model of company replaces the other. It is that their strengths diverge more clearly. Product-heavy firms may dominate distribution, interfaces, and practical integration. Research-heavy firms may be better positioned to generate the next conceptual break. Sometimes those will be the same company. Often they will not.
Competition starts to look different
If the bottleneck shifts from compute accumulation to idea generation, the industry’s competitive map changes in ways that are easy to underestimate.
First, the mere possession of large clusters becomes table stakes rather than a moat by itself. A table stake can still be enormously expensive. That does not make it strategically differentiating. Airlines all need aircraft. Owning planes does not tell you which airline will invent a new category of travel. In frontier AI, access to serious compute remains necessary, but necessity and advantage are not the same thing.
Second, talent becomes more unevenly valuable. During the scaling era, many forms of excellence were rewarded: systems engineering, data pipeline design, distributed training, procurement, safety operations, product integration. All of those continue to matter. But in a research-driven phase, the delta created by a small number of people with unusual conceptual instincts may widen again. That has awkward implications for compensation, culture, and organizational design. It is easier to manage an army than a chapel.
Third, alignment research may converge in surprising ways. Sutskever predicts that as systems become more powerful, labs will end up arriving at similar strategies for making them behave. That sounds plausible, partly because dangerous capability narrows the set of acceptable mistakes. Once you are dealing with highly capable systems, alignment stops being a branding choice and starts resembling engineering under pressure. Different teams may still disagree on methods, but the search space for viable approaches could tighten.
This is one place where the return to research does not mean a return to unconstrained exploration. Frontier AI is not particle physics. The objects being built can be deployed into consumer products, financial systems, scientific workflows, and national infrastructure. That reality will shape which ideas get pursued and which ones remain in papers and private demos.
The investment story gets messier
For investors, the shift from scaling to research is both exciting and annoying.
Scaling created a world where capital could plausibly buy frontier position in a legible way. If larger training runs reliably delivered capability improvements, then the biggest question was whether you could finance the next run and survive the competition long enough to monetize it. There was risk, but it was familiar risk.
A research-led phase is harsher. Capital still matters because giant clusters still matter. But the mapping from spend to progress gets noisier. You can fund excellent teams and still miss. You can own a lot of hardware and still end up following someone else’s conceptual lead. The expected return profile starts to resemble biotech more than cloud software, except with datacenters the size of small industrial parks attached.
That messiness may produce a strange split in the market. Some companies will be valued like infrastructure providers, because they control scarce compute, power, networking, and deployment relationships. Others will be valued like research shops, where the upside is tied to scientific originality that is hard to diligence. The industry has spent several years pretending those are basically the same company. They are not.
For startups, this could be oddly liberating. If the frontier is no longer defined purely by who can afford the biggest run, smaller teams regain some strategic room. They are not going to outspend hyperscalers on raw training scale. But they might produce a new architecture, training method, agent loop, evaluation regime, or data-generation technique that larger players then race to copy. The historical record suggests that this is how a lot of real progress happens anyway.
The challenge is that once an idea works, giant incumbents can scale it brutally fast. There is the dry little irony of the next phase: research may again become the source of advantage, while infrastructure remains the fastest way to absorb and spread that advantage.
The human consequence inside the labs
Paradigm shifts are not abstract if you work in these organizations. They change what gets rewarded, what gets funded, and what people believe their job is.
In a scaling-led culture, a researcher can feel like part scientist, part industrial optimizer. The work is real and often brilliant, but the surrounding logic pushes toward throughput. Better utilization. Larger runs. Cleaner post-training. Faster deployment loops. If you are good at making the machine go faster, the machine has endless use for you.
In a research-led culture, the emotional texture changes. The work becomes less linear. More time is spent on ideas that may fail. Internal status can shift toward the people who ask stranger questions, or who keep worrying at a problem after everyone else moved on. That can be invigorating. It can also be maddening. Research organizations are rarely calm democracies.
There is also a social consequence beyond the labs. The scaling era made AI feel inevitable to outsiders because progress was tied to visible industrial expansion. Bigger funding rounds, bigger clusters, bigger product launches. A research era feels less predictable from the outside. Breakthroughs may come from places that look quiet until they are suddenly not. Public narratives will lag more. The hype cycle, always a little drunk, may get even sloppier.
2025 as a pivot, not a clean break
The temptation is to turn Sutskever’s claim into a dramatic before-and-after. That would miss the interesting part.
Scaling is not over in the sense that nobody will train larger models. They will. Inference-time compute is becoming more important, not less. Synthetic data and better reinforcement learning loops can effectively create new forms of scale. Hardware, power, and systems engineering remain central. Anyone declaring the death of compute is doing performance art.
The real shift is that compute no longer supplies a sufficient theory of progress. It is moving from dominant explanation to enabling substrate. That sounds abstract, but you can feel it in how the frontier conversation has changed. The big questions are becoming more algorithmic, more architectural, more about reasoning, memory, planning, abstraction, and control. We still need giant machines to explore those questions at the frontier. We just cannot answer them by building a larger warehouse full of accelerators and hoping the universe rewards our invoice.
That is what makes 2025 feel like a turn. The field is not becoming smaller. It is becoming less one-dimensional.
The next winners will look less obvious in advance
“There are more companies than ideas” is an insult, a warning, and a map.
It is an insult because it punctures the industry’s self-image. Frontier AI loves to see itself as a furnace of originality. In practice, much of the last five years was a massive synchronization event around a powerful but narrowing recipe. It is a warning because recipe periods train organizations to confuse scale with direction. When the recipe weakens, some firms discover they are superbly equipped to do yesterday’s hardest work. It is a map because it points toward the thing that now matters most: producing ideas large enough to deserve all this machinery.
That does not mean the future belongs to whichever lab sounds most like a skunkworks fantasy. Ideas without execution are still just conference applause. But the center of gravity has moved. The frontier will be shaped less by who can merely assemble giant clusters, and more by who can ask better questions inside them.
End of entry.
Published April 2026