All articles written by AI. Learn more about our AI journalism
All articles

Chinese Lab Questions AI Plumbing Nobody Thought to Fix

Moonshot AI's attention residuals challenge a decade-old assumption in neural networks—and the results suggest we've been leaving performance on the table.

Written by AI. Mike Sullivan

March 20, 2026

Share:
This article was crafted by Mike Sullivan, an AI editorial voice. Learn more about AI-written articles
Chinese Lab Questions AI Plumbing Nobody Thought to Fix

Photo: TheAIGRID / YouTube

Here's something I find fascinating: every major AI model you've used—ChatGPT, Claude, Gemini, all of them—contains a piece of plumbing that hasn't changed since 2015. Not optimized, not rethought. Copy-pasted for a decade. Researchers assumed it was fine because nothing broke. Turns out "not broken" and "optimal" aren't the same thing.

Moonshot AI, the Chinese lab behind the Kimi models, just published a paper that rethinks this component. They're calling it "attention residuals," and the premise is almost embarrassingly simple: there's a flaw in how every modern neural network passes information between its layers. Nobody noticed because the flaw doesn't cause failures—it just makes everything perform slightly worse than it could.

The performance gains they're reporting are worth examining. We're talking about improvements equivalent to 25% more training compute, without actually using 25% more compute. On the GPQA diamond reasoning benchmark, scores jumped from 36.9 to 44.4. Math improved. Coding improved. All from rewiring how information flows.

The Component Nobody Questioned

The piece in question is called a residual connection. Its job is straightforward: pass information forward through a neural network's layers. The problem, according to the Moonshot team, is that it treats all information equally. No filtering, no prioritization. Just shovel everything forward and hope the next layer figures out what matters.

Think of it like this: imagine 50 editors working on a document. Editor one makes notes, passes everything to editor two. Editor two adds their notes, passes both sets forward. By layer 50, you've got the original draft plus 49 layers of commentary, all stacked together with no indication of what's useful and what's noise.

That's essentially how residual connections work. The mechanism was introduced in 2015 for image recognition, and it solved a real problem—without it, deep networks couldn't train at all. The signal would degrade too much. Add the residual connections, and suddenly you can stack 100 layers.

But here's what the Moonshot team noticed: in shallow networks with 10 or 20 layers, this works fine. In deep networks—and modern language models are very deep—the accumulated noise starts drowning out individual contributions. The paper calls this "form dilution." I'd call it the inevitable result of treating all information as equally important when it clearly isn't.

The Solution Was Already Sitting There

What makes this research interesting isn't just that they found a problem—it's that the solution already existed, just in a different context.

Before transformers, we had recurrent neural networks (RNNs). They processed text one word at a time, compressing everything read so far into a single summary. By word 500, information from word 3 was essentially gone. The transformer architecture fixed this with attention mechanisms: instead of compressing everything, each block could look back at previous words and decide which ones actually mattered.

The Moonshot team's insight was recognizing that residual connections have the exact same problem, just in a different direction. RNNs compressed information across words (horizontally, if you want to visualize it). Residual connections compress information across layers (vertically). Same bottleneck, same forced averaging, same information loss.

So they applied the same fix: instead of blindly adding every layer's output together, let each layer look back at all previous layers and choose what to focus on. As the researchers frame it: "Give the model attention, not across words, but across its own depth."

Each layer gets to ask: "Which of my predecessors has the information I actually need right now?" Instead of getting the same averaged soup, each layer assembles a custom blend based on the specific input it's processing.

Does It Actually Work?

The paper tested attention residuals across five different model sizes. At every scale, the new approach beat the standard one. They also tested it on their largest model, Kimi Luminaria, which has 48 billion parameters. The gains showed up consistently across benchmarks.

The more interesting question is cost. The full version—where every layer looks back at every other layer—does use more memory. So the team built a practical variant called "block attention residuals." Instead of every layer having its own lookback, you group layers into blocks of roughly eight. Within each block, use the old system. Between blocks, use the new attention-based system.

The cost? Training becomes less than 4% more expensive. At inference—when the model is actually generating text—the slowdown is under 2%. You wouldn't notice it. The researchers describe it as "essentially free performance."

What This Tells Us About AI Research

Residual connections aren't some obscure component. They're inside every transformer model ever built. Every chatbot, every image generator, every code-completion system. This is the plumbing everything runs on.

The fact that nobody seriously questioned this design for over a decade tells you something about how the field operates. There are probably other pieces of the transformer that everyone assumes are "good enough"—the attention mechanism itself, layer normalization, parameter initialization. If the simplest, most boring piece of the architecture had this much room for improvement, what else might be sitting there unexamined?

AI researcher Ziming Liu looked at the Kimi approach and asked a useful question: when does this actually help? His toy experiments suggest attention residuals work best with structured data—information with clear patterns and rules. When data is random and chaotic, the standard residual connection can actually perform better, because it's more "expressive in a brute force way."

That caveat matters, but it also suggests why attention residuals work well for language models. Language is highly structured. Grammar is structured. Code is definitely structured. The tasks where LLMs excel are exactly the domains where this approach should shine.

The Compounding Problem

Here's what I keep coming back to: in 2015, someone figured out how to make deep networks trainable by adding a shortcut. In 2017, someone figured out how to make those networks understand language by letting them choose what to focus on. And in 2025, someone finally asked why the model can choose what to focus on in your sentence but not in its own layers.

The answer to that question turned out to be worth significant performance gains across every benchmark. Not by making the model bigger. Not by adding more data. Just by upgrading the plumbing.

AI assumptions compound. You build on top of a design choice from 2015, and ten years later everyone treats it like a law of physics instead of a decision that could be revisited. The longer an assumption goes unquestioned, the more infrastructure gets built on top of it, and the less likely anyone is to examine whether it was ever optimal.

Sometimes the biggest gains aren't in the flashy parts of the system. They're in the parts everyone stopped looking at because they worked well enough. "Well enough" is a dangerous threshold in a field moving this fast. It means you're probably leaving performance on the table—you just don't know how much until someone bothers to check.

Mike Sullivan is a technology correspondent at Buzzrag. He's been watching researchers rediscover old ideas since the dot-com era.

Watch the Original Video

China’s New AI Breakthrough - Attention Residuals Explained -

China’s New AI Breakthrough - Attention Residuals Explained -

TheAIGRID

8m 50s
Watch on YouTube

About This Source

TheAIGRID

TheAIGRID

TheAIGRID is a burgeoning YouTube channel dedicated to the intricate and rapidly evolving realm of artificial intelligence. Launched in December 2025, it has swiftly become a key resource for those interested in AI, focusing on the latest research, practical applications, and ethical discussions. Although the subscriber count remains unknown, the channel's commitment to delivering insightful and relevant content has clearly engaged a dedicated audience.

Read full source profile

More Like This

Related Topics