Scaling Law is not dead! Core executives of Gemini reveal that Google already has a disruptive key

Wallstreetcn
2025.12.20 09:10
portai
I'm PortAI, I can summarize articles.

Sebastian Borgeaud, the head of pre-training for Google's DeepMind's Gemini, revealed that significant technological innovations in long context processing efficiency and context length expansion will occur within the next year. He also mentioned interesting discoveries in attention mechanisms that could reshape research directions. Borgeaud emphasized that the Scaling Law has not disappeared but is evolving

Is Google about to have a major breakthrough?

Recently, Sebastian Borgeaud, the head of pre-training for Google DeepMind's Gemini, made a significant revelation in an interview—

In the coming year, the field of large model pre-training will see major technological innovations in two key areas: "long context processing efficiency" and "context length expansion."

At the same time, the three giants of Google Gemini—Jeff Dean, Oriol Vinyals, and Noam Shazeer—rarely appeared together on stage, and their discussions showed an astonishing consistency with Sebastian's content.

Many forward-thinking ideas, shining with the light of wisdom, provoke deep thought.

No wonder Google remains that giant.

Google Executive Excitedly Predicts: Core Secrets of Large Models Have Been Unlocked

Sebastian Borgeaud, the head of pre-training for Google DeepMind's Gemini, stated in a recent interview that significant innovations in pre-training technology aimed at improving long context processing efficiency and further expanding model context length are expected in the coming year.

He also revealed that they have made some very interesting discoveries in attention mechanisms recently, which may reshape their research direction in the coming months.

He expressed great excitement about this.

Moreover, he made a striking statement: Scaling Law has not perished; it is merely evolving!

Sebastian Borgeaud is the head of pre-training for Gemini 3.

This is his first blog interview, providing us with an in-depth understanding of the laboratory thinking behind Gemini 3—where the changes are happening and why the current work is no longer about "training models," but about building a complete system.

Behind Gemini 3: The Future of AI is Accelerating Towards Us

After an unexpected leap, a system was born.

"If I am honest with myself, I think... we have come further than I thought we could."

Sitting in front of the microphone, Sebastian Borgeaud's tone was calm, but this statement was like a stone thrown into a lake, creating infinite ripples.

Why has Gemini 3 achieved such a significant performance leap?

Sebastian's answer seems simple: "better pre-training and better post-training."

However, behind this understatement lies a fundamental cognitive shift.

"We are no longer just building a model," he said slowly, "I believe that at this point, what we are really building is a system."

And this is the key to the disruptive progress of Gemini 3.

People often imagine that from one version of Gemini to the next, there are always some groundbreaking "secret weapons." But the truth revealed by Sebastian is that progress comes from the aggregation of countless subtle improvements. It comes from the "knobs" and optimizations discovered day after day by a large team.

He also pointed out the potential shift in the paradigm of AI development: in the past, it seemed we had unlimited data and could expand datasets at will; now, we are moving towards a "data-limited" system.

This means that future AI must learn to utilize limited data resources more efficiently to build more complex systems engineering.

This is the core value of model architecture research.

This shift will force the entire industry from a rough model of "great effort yields miracles" to a refined model of "meticulous craftsmanship."

The future competitive focus will no longer be "who has the largest data center," but rather "who has the most efficient learning algorithms," "who has the most elegant model architecture," and "who can extract more wisdom from limited data."

The brain architecture of Gemini 3—the Mixture of Experts (MoE) model—represents a clear path for the development of LLMs: shifting from the pursuit of mere "size" to the pursuit of "large and efficient, large and intelligent."

The Next Stop for AI—Long Contexts, Efficient Retrieval, and the "Cost Revolution"

Looking ahead to the technological frontier of the next few years, Sebastian pointed out exciting directions that will collectively shape the capabilities and forms of the next generation of AI.

1. Long Contexts: From "Short Memory" to "Massive Workbench"

The ultra-long context capability brought by Gemini 1.5 is already a revolution. Sebastian predicts that innovations in this area will continue to accelerate. "I believe that in the next year or so, we will see more innovations in making long contexts more efficient and further extending context length."

This is not just about "remembering more." Ultra-long contexts will turn the model into a true digital workbench: it can simultaneously load an entire codebase, multiple research papers, and long conversation histories, and conduct coherent analysis, reasoning, and creation within them This provides unprecedented possibilities for complex multi-step tasks, in-depth research, and creation, serving as a key infrastructure for advancing more powerful "intelligent agents."

2. The Evolution of Attention Mechanisms

Sebastian specifically mentioned that there have been "some truly interesting discoveries" regarding attention mechanisms recently, which he believes will shape a significant amount of research in the coming months. He is "personally very excited" about this.

This suggests that the attention mechanism, considered the cornerstone of current large models, still has enormous room for improvement.

More efficient, powerful, or feature-rich attention mechanisms could significantly enhance the model's understanding, reasoning, and computational efficiency from the ground up, serving as an important lever for pushing performance boundaries.

3. The Return of Retrieval: Teaching Models to "Look Up Information"

Sebastian previously led the "Retro" project, which explored enabling models to retrieve external knowledge bases during training and inference, rather than memorizing all knowledge in parameters. He believes this direction is far from outdated.

"I deeply believe that the long-term answer is to learn this capability in a differentiable manner." This means that future models may combine retrieval and reasoning more naturally, dynamically acquiring information from vast knowledge sources for thinking, rather than relying on later "patched" search tools.

This could make models more precise and timely, potentially breaking through the knowledge capacity limits imposed by parameter scale.

4. The "Revolution" of Efficiency and Cost

An increasingly prominent challenge is that, with the surge in users, the deployment and service costs of models have become crucial. Future research will no longer solely pursue peak performance but must also focus on how to make powerful models "cheap and easy to use."

I Still Don't See an End

As the interview neared its conclusion, when we shifted the focus back to this researcher standing at the forefront of the AI wave, Sebastian's response revealed a calm optimism and pure enthusiasm.

Despite discussing real issues such as data bottlenecks and cost challenges, Sebastian remains confident about the overall progress of AI.

"I have mentioned this many times, but there are indeed so many different things that will compound together, and there are many directions with room for improvement. I really don't see any end to this kind of work stopping us from making progress."

This confidence is not blind. It stems from what he has witnessed as a frontline navigator: a wealth of subtle yet certain areas for improvement, as well as the untapped creative vitality of the entire field In his view, at least in the coming years, this momentum of progress will not slow down.

Scale is No Longer a Myth: Noam Shazeer's First Cold Water After Returning

Next, there was a discussion among three big names: Jeff Dean, Noam Shazeer, and Oriol Vinyals.

During the meeting, we could clearly feel that Noam is no longer the radical who floors the accelerator.

He hardly initiates discussions about "disruption" and rarely uses those grandiose terms that excite people. In fact, he unusually talked multiple times about the pace of research and development, system stability, and how to operate in the long term.

As the pioneer of the Transformer, he once led the development of large models from scratch; now, as large models are advancing rapidly, he has stopped to warn: the answers provided by large models come too quickly, and the frequency of self-checking is too low.

He has repeatedly emphasized in public:

Current models are not lacking in "intelligence," but in the ability to think continuously and to repeatedly correct themselves in complex tasks.

In other words, the scale of large models is still important, but it is no longer the only variable that determines everything.

When Noam discusses reasoning, he clearly shifts the focus from "can it be stronger" to "can it be more stable."

This is the first time he has shown this judgment in public since returning to Google.

It does not sound like a denial of the past, but more like an acknowledgment of a fact: the path that purely relies on stacking parameters to keep moving forward is nearing its limits.

Consensus Among Top Scientists: A High Score Cannot Define the Future of AI

Noam mentioned a term: Slow Thinking.

It is not just about simply slowing down the pace of research and development, but about repeatedly questioning whether it is worth it, whether it is expensive, and whether it can be scaled and replicated.

At this point, intelligence is no longer an abstract capability, but an engineering expense that requires long-term investment, needing to be discussed alongside CPU and hard drives in the accounting books.

Once this step is taken, many past standards for measuring AI, such as Benchmark, are slowly becoming ineffective.

Throughout the meeting, the three top scientists did not show excitement about rankings, nor did they define progress with "who won."

Rankings are good at measuring instantaneous performance but struggle to answer long-term questions like "can it run continuously."

What Noam and Jeff repeatedly emphasized is precisely the latter: whether the model is reliable, whether it has transferability, and whether it can continuously self-correct in complex tasks.

These capabilities are difficult to compress into a neat number.

Gemini is Treated as a System, Not a Model

In this discussion, the term "System" appeared frequently, which is clearly not a rhetorical device Noam and Jeff deliberately avoided the term "a stronger model" when describing Gemini, instead repeatedly emphasizing that it is a "system" that can operate long-term and iterate continuously.

"System" and "model" may sound similar, but the underlying logic is vastly different.

A "model" is more like a one-time achievement, a momentary performance in the laboratory; whereas a "system" resembles infrastructure, focusing on stability, scalability, and the ability to quickly fix errors and continue functioning.

For someone like Noam, who places extreme emphasis on architectural design, development pace, and engineering constraints, the ability of a system to operate robustly for ten or twenty years is naturally more important than the speed of a single response.

In their view, Gemini is not a trophy for the victor, but a form of intelligence that is "long-term usable."

Because of this, the entire conversation did not rush into product promotion or engage in a tense "benchmarking against a model."

It was more about emphasizing externally: Google is not pursuing a fleeting finished product, but rather a reusable and continuously evolving intelligent industrial system.

Risk Warning and Disclaimer

The market has risks, and investment requires caution. This article does not constitute personal investment advice and does not take into account the specific investment goals, financial situation, or needs of individual users. Users should consider whether any opinions, views, or conclusions in this article align with their specific circumstances. Investing based on this is at one's own risk