Latent Space: The AI Engineer Podcast cover image

MPT-7B and The Beginning of Context=Infinity — with Jonathan Frankle and Abhinav Venigalla of MosaicML

Latent Space: The AI Engineer Podcast

00:00

Why Sparsity is not a focus for Mosaic

Sparsity is not a focus for Mosaic/nThere is currently no hardware that can accelerate Sparsity efficiently/nCerebrus has developed architecture for Sparsity/nThe field of Sparsity is open if there is hardware that can speed it up and provide cost savings/nAdding structure to Sparsity reduces the quality of resulting models/nModels love unstructured Sparsity/nPopular models always run fast on today's hardware/nNew architectures need to co-evolve with new hardware for efficient Sparsity/nTransformers and GPUs are a well-matched pair/nTPUs were designed for Transformers

Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner
Get the app