
Rethinking Model Size: Train Large, Then Compress with Joseph Gonzalez - #378
The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence)
00:00
Optimizing Model Performance with Serverless Computing
This chapter explores the intersection of economics and computing in serverless environments, focusing on strategies for efficient model training and experimentation. Key techniques such as weight pruning and quantization are discussed to enhance model efficiency and reduce costs while scaling larger machine learning models.
Transcript
Play full episode