Latent Space: The AI Engineer Podcast cover image

FlashAttention 2: making Transformers 800% faster w/o approximation - with Tri Dao of Together AI

Latent Space: The AI Engineer Podcast

CHAPTER

Exploring Alternatives to Transformer Architectures in Language Modeling

This chapter investigates potential alternatives to transformer architectures in language modeling, focusing on methods like state-based and recurrent neural networks. The discussion highlights the need to evaluate these alternatives for their efficiency and performance in processing long sequences.

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner