5min chapter

Towards Data Science cover image

120. Liam Fedus and Barrett Zoph - AI scaling with mixture of expert models

Towards Data Science

CHAPTER

Training in Stability in Distributed Transformers

Experts introduce a new communication primitive called alled to all, which isn't seen in normal like standard dente distributed transformers. So for instance, if an input is coming in am at like, time step t for the model, it might get a expert i. But now lets say we're looking at t plus one, so the prameters ted via the greating up date, and the same examples, coming in, now it might actually get sent to expert j. An expert j might have a very different output. And so you sort of fundamentally have these like a kind of very discontinuous systems. It seems like this is, again, highly speculative, but could be an

00:00

Get the Snipd
podcast app

Unlock the knowledge in podcasts with the podcast player of the future.
App store bannerPlay store banner

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode

Save any
moment

Hear something you like? Tap your headphones to save it with AI-generated key takeaways

Share
& Export

Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode