Serverless LLMs and Agentic AI with Modal โ€“ Lesson 4

BrainOmega ยท Beginner ยท๐Ÿง  Large Language Models ยท1mo ago
๐Ÿ’– Support BrainOmega โ˜• Buy Me a Coffee: https://buymeacoffee.com/brainomega ๐Ÿ’ณ Stripe: https://buy.stripe.com/aFa00i6XF7jSbfS9T218c00 ๐Ÿ’ฐ PayPal: https://paypal.me/farhadrh ๐ŸŽฅ In this video, we continue our Serverless LLMs and Agentic AI course with Lesson 4: GPUs + CPU/Memory Reservations in Modal. After learning how to define environments with custom images and control scaling behavior, we now focus on performance and hardwareโ€”specifically how resource choices directly impact model inference speed. In this lesson, youโ€™ll learn how to run the same embedding workload on default CPU resouโ€ฆ
Watch on YouTube โ†— (saves to browser)
5 Levels of AI Agents - From Simple LLM Calls to Multi-Agent Systems
Next Up
5 Levels of AI Agents - From Simple LLM Calls to Multi-Agent Systems
Dave Ebbelaar (LLM Eng)