Join Clayton Coleman, a core contributor to Kubernetes and OpenShift architect, alongside Rob Shaw, Engineering Director at Red Hat and vLLM contributor. They dive into deploying large language models (LLMs) on Kubernetes, discussing unique challenges and performance optimizations. Expect insights on the future of AI models, the pivotal role of collaborative open-source communities, and innovations like the Inference Gateway that drive efficiency in processing workloads. Get ready for an enlightening take on AI in the cloud-native space!