AI Summer cover image

AI Summer

Nathan Labenz on the future of AI scaling

Jan 27, 2025
Nathan Labenz, host of the Cognitive Revolution podcast and an AI scout, joins to discuss the recent slowdown in AI scaling. He notes that while technology adoption has lagged, significant advancements still occur in model capabilities. Labenz anticipates continued rapid progress, maintaining that we're still on the steep part of the scaling curve. The conversation also highlights AI's potential to discover new scientific concepts, emphasizing the need for a deeper understanding of scaling laws and the complexities within AI organizations.
01:18:56

Episode guests

Podcast summary created with Snipd AI

Quick takeaways

  • Despite a slowdown in AI adoption over the past 18 months, significant advancements in model capabilities remain evident across various dimensions.
  • The scaling curve in AI still indicates potential for rapid progress, although unexpected challenges may arise in development and implementation.

Deep dives

The Current State of AI Model Scaling

AI model scaling has experienced mixed outcomes since the launch of GPT-4, with the anticipated release of GPT-5 remaining elusive. Despite the introduction of new models like Google’s Gemini 1.5 Pro and Anthropic’s Claude Sonnet 3.5, the expected scaling improvements in larger versions have not materialized. This has led to speculation about underlying challenges that AI labs may be facing, as highlighted by comments from key executives in the AI field. The notion that scaling laws may not produce consistent improvements or that technical hurdles impede progress raises important questions about the future landscape of AI development.

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner