GEPA with Lakshya A. Agrawal - Weaviate Podcast #127!

Weaviate vector database · Advanced ·📄 Research Papers Explained ·9mo ago
Lakshya A. Agrawal is a Ph.D. student at U.C. Berkeley! Lakshya has lead the research behind GEPA, one of the newest innovations in DSPy and the use of Large Language Models as Optimizers! GEPA makes three key innovations on how exactly we use LLMs to propose prompts for LLMs, (1) Pareto-Optimal Candidate Selection, (2) Reflective Prompt Mutation, and (3) System-Aware Merging. The podcast discusses all of these details further, as well as topics such as Test-Time Training and the LangProBe benchmarks used in the paper! I hope you find the podcast useful! Links: GEPA: Reflective Prompt Evolution can Outperform Reinforcement Learning: https://arxiv.org/pdf/2507.19457 Lakshya A. Agrawal: https://lakshyaaagrawal.github.io/ DSPy 3.0: https://github.com/stanfordnlp/dspy/releases/tag/3.0.0 Chapters: 0:00 Welcome Lakshya! 4:50 Natural Language Rewards 11:38 Evolution of DSPy 14:55 Pareto-Optimal Candidate Selection 19:58 Domain Specific Knowledge 23:05 Exploration and Diversity 28:35 Test-Time Training 40:10 LangProBe Benchmark 48:05 Prompt Optimization and RL 52:20 Database Tuning with GEPA 1:00:00 Exciting Directions for AI
Watch on YouTube ↗ (saves to browser)
Sign in to unlock AI tutor explanation · ⚡30

Related AI Lessons

The ABCs of reading medical research and review papers these days
Learn to critically evaluate medical research papers by accepting nothing at face value, believing no one blindly, and checking everything
Medium · LLM
#1 DevLog Meta-research: I Got Tired of Tab Chaos While Reading Research Papers.
Learn to manage research paper tabs efficiently and apply meta-research techniques to improve productivity
Dev.to AI
How to Set Up a Karpathy-Style Wiki for Your Research Field
Learn to set up a Karpathy-style wiki for your research field to organize and share knowledge effectively
Medium · AI
The Non-Optimality of Scientific Knowledge: Path Dependence, Lock-In, and The Local Minimum Trap
Scientific knowledge may be stuck in a local minimum, hindering optimal progress, and understanding this concept is crucial for advancing research
ArXiv cs.AI

Chapters (11)

Welcome Lakshya!
4:50 Natural Language Rewards
11:38 Evolution of DSPy
14:55 Pareto-Optimal Candidate Selection
19:58 Domain Specific Knowledge
23:05 Exploration and Diversity
28:35 Test-Time Training
40:10 LangProBe Benchmark
48:05 Prompt Optimization and RL
52:20 Database Tuning with GEPA
1:00:00 Exciting Directions for AI
Up next
Microsoft Research Forum | Season 2, Episode 4
Microsoft Research
Watch →