Latent Space: The AI Engineer Podcast cover image

MPT-7B and The Beginning of Context=Infinity — with Jonathan Frankle and Abhinav Venigalla of MosaicML

Latent Space: The AI Engineer Podcast

NOTE

The Future of Training Efficiency

Exciting advancements in training efficiency are expected this year. The new H100s hardware from NVIDIA, along with the FP8 floating point format, are projected to provide significant improvements. By utilizing 16-bit training and now FP8, math calculations in models can be done with greater precision and cost reduction. Profiling with FP8 on H100s has already shown promising results. Additionally, architectural modifications, such as introducing sparsity, are being explored. Overall, this year promises a huge leap in cost reduction and improved performance for training.

00:00

Get the Snipd
podcast app

Unlock the knowledge in podcasts with the podcast player of the future.
App store bannerPlay store banner

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode

Save any
moment

Hear something you like? Tap your headphones to save it with AI-generated key takeaways

Share
& Export

Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode