Learn vLLM: Troubleshooting Deepseek R1 8B GPU OOM on single L4 GPU
Learn vLLM by troubleshooting errors and figuring out how to tweak the vLLM engine arguments to solve a GPU Out of Memory issue.
vLLM was deployed on K8s using KubeAI: https://github.com/substratusai/kubeai
Watch on YouTube ↗
(saves to browser)
DeepCamp AI