Latent Space: The AI Engineer Podcast cover image

MPT-7B and The Beginning of Context=Infinity — with Jonathan Frankle and Abhinav Venigalla of MosaicML

Latent Space: The AI Engineer Podcast

00:00

The Future of Training Efficiency

Exciting advancements in training efficiency are expected this year. The new H100s hardware from NVIDIA, along with the FP8 floating point format, are projected to provide significant improvements. By utilizing 16-bit training and now FP8, math calculations in models can be done with greater precision and cost reduction. Profiling with FP8 on H100s has already shown promising results. Additionally, architectural modifications, such as introducing sparsity, are being explored. Overall, this year promises a huge leap in cost reduction and improved performance for training.

Play episode from 47:57
Transcript

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app