Machine Learning Guide cover image

MLA 014 Machine Learning Hosting and Serverless Deployment

Machine Learning Guide

00:00

Using a W S Lamda to Hoast Machine Learning Models for Inference

The parallelization capabilities of a w s lamda compensates for the lack of gpu availability when running inference onh models. The minimum amount of ram that lamda provi, i think the max ram a lamda function providess three gigabites. If it's not, then you'll consider some alternatives, like batch, or some more stuff that we're going to discuss here in a bit.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app