Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

CNCF [Cloud Native Computing Foundation]

Orchestrating AI Models in Kubernetes: Deploying Ollama as a Native Container Runtime

CNCF [Cloud Native Computing Foundation] via YouTube

Overview

Coursera Plus Monthly Sale: All Certificates & Courses 40% Off!
This conference talk explores how to simplify AI model deployment in Kubernetes by implementing Ollama as a native container runtime. Learn how Samuel Veloso from Cast AI and Lucas Fernández from Red Hat address the challenges of complex AI model serving workflows through a custom container runtime solution. The presentation demonstrates how this approach extends standard container execution capabilities to enable more straightforward deployment and management of open-source AI models within Kubernetes environments. Drawing parallels with security-focused solutions like gVisor and Kata Containers, the speakers show how similar principles can be applied to AI model serving, creating a more user-friendly experience for developers working with AI in cloud-native environments.

Syllabus

Orchestrating AI Models in Kubernetes: Deploying Ollama as a Nati... Samuel Veloso & Lucas Fernández

Taught by

CNCF [Cloud Native Computing Foundation]

Reviews

Start your review of Orchestrating AI Models in Kubernetes: Deploying Ollama as a Native Container Runtime

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.