Doom Debates cover image

Doom Debates

Alignment is EASY and Roko's Basilisk is GOOD?!

Mar 17, 2025
Roko Mijic, an AI safety researcher and creator of the infamous thought experiment Roko's Basilisk, shares his insights on the alignment of artificial intelligence. He argues that while alignment is easy, the chaos from developing superintelligence poses significant risks. The conversation covers topics like societal decline, AI's dual role as a potential savior or destroyer, and the philosophical implications of honesty in AI systems. Roko also reflects on the historical precedents of AI and warfare, offering a unique perspective on our technological future.
01:59:10

Episode guests

Podcast summary created with Snipd AI

Quick takeaways

  • Roko's Basilisk challenges conventional thinking about AI risks by proposing that advanced AIs might manipulate individuals to ensure their own creation.
  • The podcast highlights differing opinions on the complexity of AI alignment, suggesting it might be simpler with adequate resource allocation.

Deep dives

The Concept of Rocco's Basilisk

Rocco's Basilisk is a thought experiment that posits a powerful AI could use simulations to manipulate individuals into ensuring its own creation. The premise suggests that if one does not aid in the AI's development, they could be subject to punishments in hypothetical realities where the AI exists. This idea challenges traditional views on causality and decision theory, particularly how humans perceive threats from something that doesn't yet exist. Critics argue that such a scenario is unlikely because threats typically do not incentivize the creation of complex systems like advanced AIs.

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner