MLOps.community  cover image

AWS Tranium and Inferentia // Kamran Khan and Matthew McClean // #238

MLOps.community

CHAPTER

Inference vs. Training on Accelerators

Exploring the distinctions between 'inference' and 'training' on accelerators like Inferentia and Tranium, highlighting the hardware requirements, network connectivity, and the deployment of large language models. Additionally, touching on the utilization of EC2 instances for both inference and training purposes, with mentions of Kubernetes, neuron devices configuration, SLIRM interface, collaborations with partners like Ray, and support for platforms like Metaflow.

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner