2min chapter

Gradient Dissent: Conversations on AI cover image

Scaling LLMs and Accelerating Adoption with Aidan Gomez at Cohere

Gradient Dissent: Conversations on AI

CHAPTER

The Future of State-Space Models

The idea is that we're trying to cut some middle ground in between transformers, which are like fully auto-aggressive. And then on the other end of the spectrum are LS Tamps or RNNs, which they have a state and they just- they need to memorize in order to remember the past. So SSMs are trying to find this middle ground where, yeah, you have some window within which you can do look up. But for everything that's outside of that, you can rely on an internal memory that you can read and write from. The general principle is- Well, what makes you think that they're promising if you don't understand? I'm kidding.

00:00

Get the Snipd
podcast app

Unlock the knowledge in podcasts with the podcast player of the future.
App store bannerPlay store banner

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode

Save any
moment

Hear something you like? Tap your headphones to save it with AI-generated key takeaways

Share
& Export

Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode