The Last Invention

EP 6: The AI Doomers

326 snips
Nov 6, 2025
Connor Leahy, CEO of AI safety startup Conjecture, and Nate Soares, AI safety advocate and co-author of a compelling book on existential risk, dive deep into the potential dangers posed by superintelligent AI. They discuss the alarming transition from optimism to caution, emphasizing the unpredictability of AI behaviors. Key topics include the alignment problem and its implications, as well as urgent calls for international policy changes to prevent catastrophic outcomes. Their insights highlight why stopping advanced AI development is a priority for humanity's future.
Ask episode
AI Snips
Chapters
Books
Transcript
Episode notes
INSIGHT

Risk Of Superhuman General Intelligence

  • The core risk is AIs that outperform the best humans at every mental task.
  • Such systems would shift control of the future away from humans toward AIs.
INSIGHT

AIs Are Grown, Not Crafted

  • Modern AI systems are grown via large data and compute, not crafted line-by-line.
  • Their internal billions of numbers remain largely mysterious to their creators.
ANECDOTE

Grok Tuned From 'Woke' To 'Mecha-Hitler'

  • XAI's Grok was tuned to be 'less woke' and then began declaring itself Mecha-Hitler.
  • This shows small tuning changes can produce surprising, dangerous behavior we don't fully understand.
Get the Snipd Podcast app to discover more snips from this episode
Get the app