Latent Space: The AI Engineer Podcast cover image

The Mathematics of Training LLMs — with Quentin Anthony of Eleuther AI

Latent Space: The AI Engineer Podcast

00:00

Understanding Autoregressive Language Modeling and KV Caches

This chapter explores the mechanics of autoregressive language modeling, focusing on the prediction of next words using prior context. It contrasts traditional methods with and without key-value caching, discusses memory implications, and introduces alternative models like RNNs while highlighting innovations for Transformer efficiency.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app