Why OpenAI Embeddings Still Work Even After You Truncate Them
Why do OpenAI embeddings still work even after you truncate a large number of dimensions?
In this video, we explore Matryoshka Representation Learning (MRL), a training technique that allows embeddings to remain useful even when you use only a prefix of the full vector. This makes it possible to trade off accuracy, memory usage, and retrieval latency at inference time.
We first look at how embeddings are normally trained using contrastive learning, and why standard embedding models do not guarantee that truncated vectors will work well. Then we see how Matryoshka Representation Learning modif…
Watch on YouTube ↗
(saves to browser)
Chapters (6)
Truncated OpenAI Embeddings Still Work
0:46
The Idea Behind Matryoshka Embeddings
2:03
How Embeddings Are Normally Trained (Contrastive Learning)
3:28
How Matryoshka Representation Learning Changes the Loss
5:16
Why Matryoshka Embeddings Are Useful for Vector Search & RAG
6:21
Results from the MRL Paper
DeepCamp AI