Рет қаралды 2,264
Today we dive into running AI models on Kubernetes with GPU support. Learn how to manage GPUs in Kubernetes clusters, create GPU nodes, and optimize resource usage without breaking the bank. We'll walk you through setting up a Google Cloud Kubernetes cluster (the same logic should apply to other Cloud providers), deploying AI models like Ollama's Llama2, and handling GPU partitioning. Watch now to master GPU-based AI workloads in Kubernetes!
#Kubernetes #GPU #AI
▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬
Sponsor: CAST AI
🔗 cast.ai
▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬
Consider joining the channel: kzbin.infojoin
▬▬▬▬▬▬ 🔗 Additional Info 🔗 ▬▬▬▬▬▬
➡ Transcript and commands: devopstoolkit.live/ai/unlock-the-power-of-gpus-in-kubernetes-for-ai-workloads
▬▬▬▬▬▬ 💰 Sponsorships 💰 ▬▬▬▬▬▬
If you are interested in sponsoring this channel, please visit devopstoolkit.live/sponsor for more information. Alternatively, feel free to contact me over Twitter or LinkedIn (see below).
▬▬▬▬▬▬ 👋 Contact me 👋 ▬▬▬▬▬▬
➡ Twitter: vfarcic
➡ LinkedIn: www.linkedin.com/in/viktorfarcic/
▬▬▬▬▬▬ 🚀 Other Channels 🚀 ▬▬▬▬▬▬
🎤 Podcast: www.devopsparadox.com/
💬 Live streams: kzbin.info
▬▬▬▬▬▬ ⏱ Timecodes ⏱ ▬▬▬▬▬▬
00:00 AI Inference with GPUs
01:30 CAST AI (sponsor)
02:29 Using GPUs for AI Inference in Kubernetes