Latent Space: The AI Engineer Podcast cover image

FlashAttention 2: making Transformers 800% faster w/o approximation - with Tri Dao of Together AI

Latent Space: The AI Engineer Podcast

00:00

Exploring Alternatives to Transformer Architectures in Language Modeling

This chapter investigates potential alternatives to transformer architectures in language modeling, focusing on methods like state-based and recurrent neural networks. The discussion highlights the need to evaluate these alternatives for their efficiency and performance in processing long sequences.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app