Hierarchical Memory: Context Management in Agents — Sally-Ann Delucia
Skills:
Agent Foundations90%
The naive solution is truncation. The obvious solution is summarization. Neither worked — and the Arize team found out the hard way while building an AI agent that had to analyze the very trace data it was generating.
A year of lessons from building Alyx, starting with the vicious loop that defined the problem: Alex runs on trace data, the spans grow, the context limit hits, it fails and tries again. The talk covers why truncation breaks reasoning, why summarization gives the LLM too much control, and how head/tail preservation with a retrievable memory store is what actually held. Then: long session evals, sub-agents as the answer when one context accumulates too much, and what they found when they went looking for secrets in the Claude Code source release.
Speaker info:
- https://www.linkedin.com/in/sallyann-delucia-59a381172/
- related links
Watch on YouTube ↗
(saves to browser)
Sign in to unlock AI tutor explanation · ⚡30
More on: Agent Foundations
View skill →Related AI Lessons
⚡
⚡
⚡
⚡
Understanding Real-Time Customer Intent: The New Frontier for Retail AI Chatbots
Medium · AI
Artificial Intelligence Is Not Replacing Humans - It’s Replacing Certain Behaviors
Medium · AI
How I cut my LangChain agent's token costs by 93% with one import
Dev.to · Mahika jadhav
5 Passive Income Streams Your AI Agent Can Run While You Sleep
Dev.to AI
🎓
Tutor Explanation
DeepCamp AI