Lex Fridman Podcast cover image

#490 – State of AI in 2026: LLMs, Coding, Scaling Laws, China, Agents, GPUs, AGI

Lex Fridman Podcast

00:00

Architectural Tweaks: MOE, Attention Variants, and Efficiency

Sebastian explains mixture-of-experts, attention tweaks (group query, sliding window) and KV-cache trade-offs across models.

Play episode from 48:50
Transcript

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app