Intelligent Machines (Audio)

IM 844: Poob Has It For You - Spiky Superintelligence vs. Generality

28 snips
Nov 6, 2025
Jeremy Berman, a post-training researcher at Reflection.ai, dives into the evolution of AI's capabilities. He addresses the challenge of 'spiky superintelligence'—smart in specific tasks but limited overall. The discussion contrasts pre-training and post-training methods, underscoring the potential of reinforcement learning for developing generality in AI. Berman shares insights on ARC AGI benchmarks, the risk of AGI, and why open-weight models are crucial for adoption. Their outlook on how AI could learn tasks easier for humans promises a fascinating future.
Ask episode
AI Snips
Chapters
Books
Transcript
Episode notes
INSIGHT

Pre- vs Post-Training Distinction

  • Pre-training compresses internet knowledge into models but leaves them as document completers.
  • Post-training is required to make models useful for human tasks and personalities.
INSIGHT

Why ARC Exposes LLM Weaknesses

  • ARC puzzles revealed that LLMs trained only to predict tokens fail at novel reasoning tasks.
  • Solving such puzzles requires models to generate new solutions, not memorize web patterns.
ANECDOTE

Iterative Code Evolution Won ARC Briefly

  • Jeremy described his ARC 2024 entry: he generated many Python programs and iteratively refined the best ones.
  • His approach won briefly until reinforcement-learned models outperformed it.
Get the Snipd Podcast app to discover more snips from this episode
Get the app