Latent Space: The AI Engineer Podcast cover image

MPT-7B and The Beginning of Context=Infinity — with Jonathan Frankle and Abhinav Venigalla of MosaicML

Latent Space: The AI Engineer Podcast

00:00

The Future of Training Efficiency

Exciting advancements in training efficiency are expected this year. The new H100s hardware from NVIDIA, along with the FP8 floating point format, are projected to provide significant improvements. By utilizing 16-bit training and now FP8, math calculations in models can be done with greater precision and cost reduction. Profiling with FP8 on H100s has already shown promising results. Additionally, architectural modifications, such as introducing sparsity, are being explored. Overall, this year promises a huge leap in cost reduction and improved performance for training.

Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner
Get the app