MLOps.community  cover image

LLM Distillation and Compression // Guanhua Wang // #278

MLOps.community

00:00

Optimizing Training Through Quantization and Communication Reduction

This chapter explores the role of quantization in training machine learning models, detailing its application for improving efficiency before and after communication. It also introduces a novel framework designed to minimize communication overhead during model training, focusing on the trade-offs of bit levels in quantization.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app