The MAD Podcast with Matt Turck

The End of GPU Scaling? Compute & The Agent Era — Tim Dettmers (Ai2) & Dan Fu (Together AI)

6 snips
Jan 22, 2026
Tim Dettmers, an assistant professor at Carnegie Mellon University, and Dan Fu, an assistant professor at UC San Diego, dive deep into the future of AGI. They debate the limitations of current hardware versus the untapped potential of efficient utilization. Tim warns of physical constraints like the von Neumann bottleneck, while Dan emphasizes better performance through optimized kernels. The conversation also reveals how agents can enhance productivity, with practical advice on leveraging them effectively for work automation and innovation in AI architectures.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
INSIGHT

Physical Limits Shape Model Progress

  • Computation is constrained by physical memory movement and the von Neumann bottleneck.
  • Architectural and hardware limits create diminishing returns for raw GPU improvements.
INSIGHT

Models Lag Behind Hardware Reality

  • Many current models were trained on older clusters and underutilized hardware.
  • Dan estimates up to ~100x more effective compute is already available when accounting for new chips and utilization gains.
INSIGHT

Pretraining vs Post-Training Tradeoff

  • Pre-training builds general capability while post-training tailors useful, product-ready skills.
  • Post-training and product feedback matter more for real-world usefulness than raw scale alone.
Get the Snipd Podcast app to discover more snips from this episode
Get the app