The MLOps Podcast cover image

🎨 Stable Diffusion and generative models with David Marx

The MLOps Podcast

00:00

Do You Have Any Counterintuitive Challenges With GPU Inference?

Inference is in any shape or form using a GPU. One of the things that differentiates how we do inference kind of at like our production level versus how you will see it on non-GPU servers, right? The time component for not just the inference, but even just spinning things up takes a lot longer. I think that really annoys more conventional back end developers too.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app