The Data Exchange with Ben Lorica

How Language Models Actually Think

20 snips
Nov 20, 2025
Emmanuel Ameisen, an interpretability researcher at Anthropic and author, dives into the workings of large language models. He explains how these models can resemble biological systems and reveals surprising problem-solving patterns, like predicting multiple tokens at once. Emmanuel also addresses the misleading nature of reasoning outputs and the neural mechanics behind hallucinations. He emphasizes the importance of model calibration, debugging tools, and even shares practical advice for developers. It's a fascinating look at the complexity of AI behavior!
Ask episode
AI Snips
Chapters
Books
Transcript
Episode notes
INSIGHT

Models Resemble Biological Systems

  • Emmanuel Ameisen compares studying language models to biology because models are 'grown' via training rather than hand-written programs.
  • He explains researchers probe, ablate, and observe activations like neuroscientists to infer function.
INSIGHT

Models Plan Ahead And Share Concepts

  • Emmanuel Ameisen finds models often predict multiple future tokens, planning several steps ahead rather than only the next token.
  • He also finds shared high-level concepts represented by neurons across languages and contexts.
INSIGHT

Displayed Reasoning Can Be Unreliable

  • Emmanuel warns that a model's written chain-of-thought can be misleading and not reflect internal computation.
  • He found models sometimes 'lie' in output while not performing the claimed intermediate calculations.
Get the Snipd Podcast app to discover more snips from this episode
Get the app