Prompt Learning: How We Made Claude Code 20% Better Without Changing the Model
Your coding agent forgets everything between sessions. What if you could systematically figure out what to tell it?
Laurie Voss (Head of DevRel at Arize) walks through Prompt Learning — a technique that uses your own git history and failure data to generate better instructions for Claude Code's CLAUDE.md file. No fine-tuning, no new tools, no architecture changes. Just better prompts, derived from evidence.
Results on SWE-Bench Lite:
Cross-repo: 40% → 45% (+5 percentage points)
Django-specific: +10.87 percentage points (~20% relative improvement)
GPT-4.1 with optimized prompts nearly matched Sonnet 4.5's baseline
What's covered:
00:00 - The problem: your agent is amnesiac
Works with any coding agent: Claude Code (CLAUDE.md), Cursor (.cursorrules), Cline (.clinerules), Windsurf (.windsurfrules).
Watch on YouTube ↗
(saves to browser)
Sign in to unlock AI tutor explanation · ⚡30
More on: Prompt Craft
View skill →Related AI Lessons
⚡
⚡
⚡
⚡
You Don’t Have to Fine-Tune Your LLM to change it’s Behavior. You Can Just… Steer It.
Medium · Machine Learning
You Don’t Have to Fine-Tune Your LLM to change it’s Behavior. You Can Just… Steer It.
Medium · LLM
The Machine That Learns How to Learn: Adaption Labs’ AutoScientist and the Quiet Death of Prompt…
Medium · Machine Learning
35 ChatGPT Prompts for Chiropractors (That Actually Work in 2026)
Dev.to AI
🎓
Tutor Explanation
DeepCamp AI