This 24-minute conference talk from DevConf.IN 2025, presented by Swaraj Pande, explores RamaLama, an innovative tool that bridges AI and containerization technologies. Discover how RamaLama simplifies AI model deployment through containerized environments, offering privacy-focused, GPU-optimized workflows for seamless model serving. Learn about practical applications including Podman Quadlet and Kubernetes YAML generation that can transform both local and collaborative AI model management. The presentation highlights RamaLama's compatibility with multiple runtimes and popular registries such as Hugging Face, Ollama, and OCI, demonstrating how it streamlines AI workflows to make working with AI models straightforward and efficient. Access accompanying slides and additional resources through the provided link to enhance understanding of this powerful containerization solution for AI deployment.
Overview
Syllabus
Who Needs Drama When You Have RamaLama? - DevConf.IN 2025
Taught by
DevConf