Lex Fridman Podcast cover image

#459 – DeepSeek, China, OpenAI, NVIDIA, xAI, TSMC, Stargate, and AI Megaclusters

Lex Fridman Podcast

00:00

LLM Token Pricing

  • Input tokens are cheaper than output tokens in LLMs like OpenAI because they can be processed in parallel.
  • Output tokens are generated autoregressively, requiring sequential processing and reading/updating the entire key-value cache for each token.
Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app