The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence) cover image

Language Modeling With State Space Models with Dan Fu - #630

The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence)

00:00

Exploring Context Length and Architectural Innovations in Language Models

This chapter investigates the transferability of state space models in language processing, particularly their capabilities with longer context lengths. It also introduces the innovative 'Hyena' approach for improving performance through restructured convolutional methods and discusses the future of model architectures.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app