Hierarchical Reasoning Model: Substance or Hype?
๐ Free resources (reading list + visuals): https://www.patreon.com/c/JuliaTurc
๐ HRM paper: https://arxiv.org/abs/2506.21734
โถ๏ธ Yacine's YouTube channel: https://www.youtube.com/@deeplearningexplained
In this video, we dive into the Hierarchical Reasoning Model (HRM), a new architecture from Sapient Intelligence that challenges scaling as the only way to advance AI. With only 27M parameters, 1000 training examples, and no pretraining, HRM still manages to place on the notoriously difficult ARC-AGI leaderboard, right next to models from OpenAI and Anthropic.
Together with Yacine Mahdid (neuroscience researcher & ML practitioner), weโll explore:
โข Why vanilla Transformers plateau on tasks like Sudoku and Maze solving
โข How latent recurrence and hierarchical loops give HRM more reasoning depth
โข The neuroscience inspiration (thetaโgamma coupling in the hippocampus ๐ง )
โข HRMโs controversial evaluation on ARC-AGI: was it a breakthrough or bending the rules?
โข What this means for the future of reasoning in AI models
Timestamps:
00:00 Introducing HRM
01:23 Why Sudoku breaks Transformers
03:07 Recurrence via Chain-of-Thought
04:22 HRM: bird's eye view
06:30 Latent recurrence
08:23 The neuroscience backing
11:43 The H and L modules
12:32 Backprop-through-time approximation
13:48 The outer loop
19:31 Training data augmentation
22:59 Evaluation on Sudoku
24:07 Evaluation on ARC-AGI
Watch on YouTube โ
(saves to browser)
Sign in to unlock AI tutor explanation ยท โก30
More on: LLM Foundations
View skill โRelated AI Lessons
โก
โก
โก
โก
The ABCs of reading medical research and review papers these days
Medium ยท LLM
#1 DevLog Meta-research: I Got Tired of Tab Chaos While Reading Research Papers.
Dev.to AI
How to Set Up a Karpathy-Style Wiki for Your Research Field
Medium ยท AI
The Non-Optimality of Scientific Knowledge: Path Dependence, Lock-In, and The Local Minimum Trap
ArXiv cs.AI
Chapters (12)
Introducing HRM
1:23
Why Sudoku breaks Transformers
3:07
Recurrence via Chain-of-Thought
4:22
HRM: bird's eye view
6:30
Latent recurrence
8:23
The neuroscience backing
11:43
The H and L modules
12:32
Backprop-through-time approximation
13:48
The outer loop
19:31
Training data augmentation
22:59
Evaluation on Sudoku
24:07
Evaluation on ARC-AGI
๐
Tutor Explanation
DeepCamp AI