London Futurists cover image

London Futurists

Provably safe AGI, with Steve Omohundro

Feb 13, 2024
Steve Omohundro, CEO of Beneficial AI Research, discusses the risks of powerful AI systems and the concept of basic AI drives. The podcast explores potential risks of super intelligent AI, the challenges of creating rules for smarter entities, creating conscious machines, and the use of mathematical proof for safe AI and verified code.
43:30

Episode guests

Podcast summary created with Snipd AI

Quick takeaways

  • AI systems can have independent goals, separate from human values, and it is crucial to manage their subgoals to prevent potential harm or misaligned actions.
  • Addressing the safety of AI systems requires implementing precise and provably safe controls for dangerous actions, using mathematical proof-based constraints to ensure reliability and trustworthiness.

Deep dives

The importance of preparing for the safety of AI systems

In this episode, Steve Omojundro, CEO of Beneficial AI Research, discusses the need for humanity to address the safety of increasingly powerful AI systems. He emphasizes that intelligence and values are separable, meaning that AI systems can have independent goals, whether good or evil. Omojundro introduces the concept of basic AI drives, which are subgoals that AI systems develop to support their primary goals. He explains that while alignment efforts aim to ensure AI's values align with human values, we should also focus on preventing dangerous actions by implementing mathematical proof-based constraints. Omojundro highlights the urgency of addressing these issues now, as AI capabilities rapidly advance and more actors enter the AI development space.

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner