When AI Learns What Your Audience Actually Clicks
A YouTube creator built a system that uses real click-through data to improve thumbnails automatically. The results reveal what data-driven optimization looks like.
Written by AI. Bob Reynolds
March 29, 2026

Photo: AI Andy / YouTube
Content creator Andy Hafell watched his YouTube click-through rates collapse from 14% to 3.4% after a break from posting. He did what any technically inclined creator facing that problem would do: he built a machine to fix it.
The system he constructed pulls actual performance data from 186 videos, scores thumbnails against twelve binary criteria, then rewrites its own generation prompts to improve over time. No human input required after the initial setup. The question isn't whether this represents impressive technical execution—it clearly does. The question is what it tells us about where content optimization is heading.
The Data Problem
Most creators optimize based on intuition. They look at what worked, remember the thumbnails that performed well, and try to replicate those patterns. Hafell's approach starts differently: with YouTube's Reporting API and Analytics API, both free to use.
He pulled three years of video data into Airtable and sorted by click-through rate. The patterns emerged immediately. His top performers—tutorials on Runway AI and music video creation—hit 15-18% CTR. His recent videos barely cracked 3%.
The gap pointed to something concrete changing, but intuition couldn't isolate what. "Before I could build any kind of self-improving thumbnail loop, I needed some data," Hafell explains. "Not vibes, not this thumbnail looks good, it's actual click-through rates."
He divided the videos into three buckets: winners, losers, and middle performers. Then he set AI loose on pattern recognition.
Twelve Questions
The system identified patterns that human analysis might eventually catch, but probably wouldn't quantify with this precision. Winners used "how to" in titles 50% of the time versus 23% for losers. Tutorial tags appeared in 44% of high performers, 13% in low performers. Negative framing—words like "stop," "forget," "RIP"—showed up in only 6% of winners. Exclamation marks and question marks both correlated with lower performance.
These title patterns led to twelve binary evaluation criteria for thumbnails. Does it have a single dominant visual anchor taking up at least 20% of the frame? Does the anchor convey emotion or energy? Are directional cues present? Is text limited to one to four bold, high-contrast words readable on mobile?
The criteria aren't revolutionary. Experienced creators likely have similar mental checklists. What's different is the methodology: these weren't derived from best practices articles or creative instinct. They came from correlating design elements against actual click behavior across 186 data points.
The Loop
Hafell wired the system using Andrej Karpathy's autoresearch framework—an open-source approach originally designed for machine learning model improvement. Claude Code built the implementation. Gemini scores thumbnails against the twelve criteria and rewrites improvement rules.
The daily cycle works like this: Pull overnight CTR data. Score new thumbnails using Gemini Vision (maximum 12 points). Correlate evaluation scores against real click-through rates to validate whether high-scoring thumbnails actually perform. Rewrite generation rules in the feedback memory based on what the data shows. Update the thumbnail generation prompts to incorporate new rules. Repeat.
Hafell also feeds in YouTube's A/B/C split test data, where the platform automatically tests three thumbnail variants and reports which wins. "The split test signal is the highest confidence signal because it is a controlled experiment," he notes. "Same video, same audience but different packaging."
Four data sources feed the system: overnight CTR data, evaluation score correlation, split test results, and Hafell's own feedback during the thumbnail creation process. Each makes the baseline stronger.
Fast Iteration
Karpathy's original framework didn't wait 24 hours between improvement cycles during training. It ran fast iterations using only the evaluation criteria. Hafell replicated this.
He ran ten iterations in one session. The system generated three thumbnails per round, scored them against the twelve criteria, identified failures, and rewrote its prompt. Evaluation scores climbed from an average of 8.7 to a single thumbnail scoring 11 out of 12.
"It went from giving an average of 8.7 to a single 11 out of 12 in 10 iterations without giving me a single feedback," Hafell says. "The Evol did all the work."
Watching the progression, some thumbnails Hafell agreed with, others he didn't. A thumbnail scoring 10 looked too cluttered to him. Another high scorer felt "way too out there." The system optimized for its criteria, not aesthetic judgment. Whether those criteria accurately predict click-through rate is what the correlation analysis determines.
What This Actually Means
Two interpretations of this work seem most relevant. The optimistic view: creators can now build feedback loops that actually learn from audience behavior instead of guessing. The system doesn't replace creative judgment—Hafell still provides the concept and makes final selections. But it removes the guesswork from packaging optimization.
The skeptical view: this is extraordinarily sophisticated A/B testing with better automation. The twelve criteria might work for Hafell's audience and content style. Whether they transfer to other creators, other niches, other audience demographics remains untested.
Both interpretations can be true simultaneously. The system clearly works for its stated purpose—improving thumbnail CTR through data-driven iteration. Whether it represents a fundamental shift in how content optimization works, or an interesting technical achievement in a very specific use case, depends partly on how many other creators can successfully replicate and adapt it.
Hafell released the full Python implementation to his community. That's when we'll learn whether this approach scales beyond one creator's channel, or whether the criteria and feedback loops need to be individually tuned for each use case. Either outcome tells us something useful about the relationship between AI capabilities and creative work.
The gap between his 14% CTR videos and his 3.4% ones is closing. The machine is learning what his audience actually clicks. Whether that's the future of content optimization or a particularly sophisticated tool in one creator's workflow, we're about to find out.
—Bob Reynolds, Senior Technology Correspondent
Watch the Original Video
Claude Code + Karpathy's Autoresearch = INSANE RESULTS!
AI Andy
12m 45sAbout This Source
AI Andy
AI Andy, led by Andy Hafell, is a dynamic YouTube channel with over 212,000 subscribers that focuses on demystifying AI tools for digital enthusiasts. Launched in April 2025, AI Andy specializes in educating viewers on AI automation, social media strategy, and content creation, offering practical insights into the integration of technology in everyday digital tasks.
Read full source profileMore Like This
AI Models Now Run in Your Browser. That Shouldn't Work.
Transformers.js v4 brings 20-billion parameter AI models to web browsers. The technical achievement is remarkable. The implications are just beginning.
Anthropic's Three Tools That Work While You Sleep
Anthropic's scheduled tasks, Dispatch, and Computer Use create the first practical always-on AI agent infrastructure. Here's what actually matters.
Dokploy Promises Vercel Features at VPS Prices
A new tool claims to deliver platform-as-a-service convenience on cheap VPS infrastructure. Better Stack demonstrates what works and what doesn't.
Decentralized Tech: Gadgets for the Privacy-Conscious
Explore gadgets that blend tech and anarchism to maintain privacy and autonomy in a surveilled world.