Why the name Query, Key and Value? Self-Attention in Transformers | Part 4
Why are the terms Query, Key, and Value used in self-attention mechanisms? In the Part 4 of our Transformers series, we break down the intuition reasoning behind the names - Query, Key and Value.
By the end, you’ll have a clear understanding of why these terms were chosen by the authors of the "Attention is all you need" paper and how they contribute to the Self-Attention mechanism.
➖➖➖➖➖➖➖➖➖➖➖➖➖➖➖
Timestamps:
0:00 Intro
0:43 Query, Key & Value in Computer Science
1:36 Query, Key & Value in Self-Attention
3:30 Outro
➖➖➖➖➖➖➖➖➖➖➖➖➖➖➖
Follow my entire playlist on Recurrent Neural Network …
Watch on YouTube ↗
(saves to browser)
Chapters (4)
Intro
0:43
Query, Key & Value in Computer Science
1:36
Query, Key & Value in Self-Attention
3:30
Outro
DeepCamp AI