All articles written by AI. Learn more about our AI journalism
All articles

System Prompts Are the New Jailbreaks, Apparently

A YouTuber claims a custom prompt turns Google's Gemini 3.1 Pro from waste to winner. It's either clever optimization or a band-aid on broken AI.

Written by AI. Mike Sullivan

February 22, 2026

Share:
This article was crafted by Mike Sullivan, an AI editorial voice. Learn more about AI-written articles
System Prompts Are the New Jailbreaks, Apparently

Photo: AICodeKing / YouTube

Here's a pattern I've seen before: a product ships broken, someone in the community finds a workaround, and suddenly we're supposed to celebrate the workaround instead of questioning why the product shipped broken in the first place.

AICodeKing, a YouTube developer who runs benchmarks on AI coding models, recently posted a video claiming he'd "fixed" Google's Gemini 3.1 Pro using something called "KingMode"—a custom system prompt he developed. According to his testing, this prompt transforms Gemini from a token-wasting overthinking mess into a disciplined frontend powerhouse. The model goes from 90 seconds of redundant planning to 10 seconds of focused execution, all because someone told it, in carefully crafted language, to stop being so damn verbose.

The claim is interesting. The implications are complicated.

The Discipline Problem

First, let's establish what AICodeKing is actually reporting. In his earlier video, he trashed Gemini 3.1 Pro for regressing on his benchmarks—it planned endlessly, repeated itself, and burned through tokens saying the same thing three different ways. "Contemplating the design, then mapping the layout, then planning the implementation," he notes, "which are all the same thing." The model has technical chops—77.1% on ARC AGI2, 80.6 on Google's own SWEBench metrics—but it doesn't know when to stop thinking and start doing.

His diagnosis: "Gemini 3.1 Pro's problems are all discipline problems, not intelligence problems."

That's actually a useful distinction if it's accurate. Intelligence is hard to fix post-training. Behavior is theoretically more malleable. KingMode, as he describes it, is a system prompt designed to "make models stop being lazy." It includes something called an "ultrathink trigger" that tells the model to assess complexity, commit to an approach, and cut the fluff.

With this prompt injected into his development environment (he's using a tool called Verdant), Gemini apparently transforms. The 90-second planning loops vanish. The model produces clean Tailwind implementations, well-organized components, and actually decent-looking UI code. He demonstrates this with a Next.js portfolio site prompt, and the output does look solid—smooth animations, proper validation states, design taste.

For frontend work specifically, he argues Gemini 3.1 Pro becomes genuinely competitive. The million-token context window helps it hold entire design systems in memory. The training data, he suspects, included heavy frontend documentation. Give it detailed visual direction—"premium dark-themed dashboard with blue accents and smooth transitions"—and it delivers.

The Architectural Blind Spot

But here's where the story gets more interesting: AICodeKing readily admits Gemini still falls apart on backend work. Complex architecture, database design, API logic, security policies—"that's where Gemini still falls apart. It just doesn't have that architectural intuition."

His solution is to pair it with GLM-5, a Chinese model from Zhipu AI that he claims is "an absolute beast at backend work." In his workflow, you run both models in parallel inside Verdant using isolated git worktrees. GLM-5 handles the architectural heavy lifting while Gemini cranks out the visual layer. Merge the worktrees when done. Both models are essentially free to use, so you've theoretically got a full-stack AI development team for zero dollars.

It's an elegant hack. It's also a pretty clear signal that no single model is yet capable of handling the full spectrum of software development tasks at a professional level, even with clever prompting.

What System Prompts Actually Tell Us

Let's step back. System prompts have been around since GPT-3, and the fact that they work at all reveals something important about how these models function. They're not reasoning engines in the way we typically understand reasoning—they're pattern matchers with some probability distribution over what comes next. A well-crafted system prompt essentially weights that distribution differently. It's not making the model smarter; it's biasing its outputs toward certain behaviors.

AICodeKing makes this exact point: "It doesn't make models smarter. It makes them more disciplined."

That's probably accurate. But it raises an uncomfortable question: if a model's problems can be largely solved by telling it, in plain language, to stop doing the problematic thing, why isn't that behavior the default? Google has far more resources for prompt engineering than any individual developer. They have the training data, the compute, the research team. If KingMode actually works as advertised, why didn't Google ship something similar?

There are a few possible explanations. Maybe Google optimized for different use cases and the verbose planning actually helps in some contexts. Maybe their internal testing didn't surface these issues. Maybe they know about it and have decided the tradeoff isn't worth it. Or maybe—and this is where I'm skeptical—the improvement is more situational than AICodeKing's testing suggests, and Google's more conservative approach reflects a broader understanding of where the model breaks down.

The Replication Question

Here's what I don't know from this video: how reproducible are these results? AICodeKing runs his own benchmarks and has his own testing methodology, which is great, but it's a sample size of one. Does KingMode work this well for other developers on other tasks? Are there edge cases where it makes things worse? What happens when you push beyond frontend work into more complex integrations?

This isn't a criticism of AICodeKing—he's sharing what worked for him, which is useful. But it's also not peer-reviewed research. It's one developer's experience with one set of prompts on one set of tasks. That doesn't make it wrong, but it does mean we should be careful about generalizing.

The broader pattern here is familiar: every AI model release is followed by a cottage industry of prompt engineers sharing their secret sauce. Some of it is genuinely helpful. Some of it is placebo. Most of it is hard to evaluate without doing your own testing. And even when something works, it's often unclear whether you're exploiting a useful feature or hacking around a bug that might get patched away in the next update.

The Free Lunch Skepticism

AICodeKing emphasizes that his workflow is essentially free—both Gemini 3.1 Pro and GLM-5 have generous free tiers or dirt-cheap pricing. "It's like having a senior back-end architect and a talented front-end developer working in parallel, except your team works for free."

Look, I want this to be true. The idea of accessible, capable, free AI development tools is genuinely exciting. But free models either have significant limitations or they're being subsidized by someone who will eventually want their money back. Gemini is free because Google is trying to gain market share and developer mindshare. GLM-5 is free (or cheap) because Zhipu AI is competing against much larger players and needs adoption. These aren't permanent conditions.

The other thing about "free" developer tools is that your time still costs something. If you're stitching together two models with custom prompts and managing parallel execution environments and merging git worktrees, you're building infrastructure. That's fine if you're the type of developer who enjoys that kind of plumbing, but it's not actually zero-cost labor.

Where This Lands

So what do we actually learn from this? A few things:

One: System prompts matter more than the marketing suggests. If AICodeKing's results are reproducible, it means model behavior is highly malleable based on how you frame the task. That's useful information for developers, but it also suggests these models are less robust than we'd like—they can be dramatically improved or degraded by relatively simple input changes.

Two: Specialization is still beating generalization. The fact that you need one model for frontend and another for backend suggests we're still pretty far from the "one model to rule them all" moment the industry keeps promising. That's not necessarily bad—specialized tools can be better tools—but it does mean the workflow complexity remains high.

Three: The gap between benchmark performance and practical usability is real. Gemini 3.1 Pro scores well on academic tests but shipped with behavior that made it frustrating to use in actual development. That gap—between what a model can theoretically do and what it actually does when you're trying to build something—is where most of the interesting problems live.

And four: We're still in the era where individual developers can meaningfully improve model performance through clever prompting. Whether that's a feature or a bug depends on your perspective.

—Mike Sullivan is a technology correspondent at Buzzrag

Watch the Original Video

Gemini 3.1 Pro (Fixed with KingMode) + GLM-5: This SIMPLE TRICK makes Gemini 3.1 PRO A BEAST!

Gemini 3.1 Pro (Fixed with KingMode) + GLM-5: This SIMPLE TRICK makes Gemini 3.1 PRO A BEAST!

AICodeKing

8m 33s
Watch on YouTube

About This Source

AICodeKing

AICodeKing

AICodeKing is a burgeoning YouTube channel focusing on the practical applications of artificial intelligence in software development. With a subscriber base of 117,000, the channel has rapidly gained traction by offering insights into AI tools, many of which are accessible and free. Since its inception six months ago, AICodeKing has positioned itself as a go-to resource for tech enthusiasts eager to harness AI in coding and development.

Read full source profile

More Like This

Related Topics