4min snip

Last Week in AI cover image

#150 - GPT Store, new Nvidia chips, DeepMind’s robotics progress, bad uses of AI

Last Week in AI

NOTE

Efficiency of State-Space Models and Mixture of Experts in AI Architecture

State-space models offer linear time inference with respect to context length, making them more computationally feasible compared to transformers. By combining state-space models with mixture of experts technique, the recent Mamba model showcased improved performance compared to both original Mamba model and transformers. This combination increases the model's power without significantly increasing computational costs, making it a promising approach for more efficient AI architecture. The integration of these two techniques could potentially lead to better models and scalability in AI architecture, providing a cost-effective solution for model development and training.

00:00

Get the Snipd
podcast app

Unlock the knowledge in podcasts with the podcast player of the future.
App store bannerPlay store banner

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode

Save any
moment

Hear something you like? Tap your headphones to save it with AI-generated key takeaways

Share
& Export

Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode