

Steve Omohundro on Provably Safe AGI
15 snips Oct 5, 2023
Steve Omohundro, co-author of Provably Safe Systems, discusses the concept of provable safety in AI, formalizing safety, provable contracts, proof-carrying code, language models' logical thinking, AI doing proofs for us, risks of totalitarianism, tamper-proof hardware, least-privilege guarantee, basic AI drives, AI agency and world models, self-improving AI, and the overhyping of AI.
Chapters
Transcript
Episode notes
1 2 3 4 5 6 7 8 9 10 11 12
Introduction
00:00 • 4min
Creating Provably Safe Systems for AGI
04:27 • 23min
Guard Rails for AI
27:13 • 19min
The Relevance of Proof Carrying Code in AI Safety
46:25 • 23min
Discovering vs Verifying Proofs
01:09:23 • 5min
Mechanistic Interpretability and AI Alignment
01:14:32 • 4min
The Potential Impact of Quantum Computing on Cryptographic Systems
01:18:27 • 2min
Avoiding a Stupid Future: Ensuring Safe Governance of Programs
01:20:01 • 2min
The Importance of Mathematical Proofs and Hardware Security in AGI
01:22:30 • 7min
Protecting Cryptographic Hardware and the Need for Tamper-Proofing
01:29:30 • 5min
Incentives, Security, and Risk Mitigation in AI
01:34:11 • 12min
AI Agency and the Implications for Humanity
01:46:19 • 16min