AI Safety Fundamentals cover image

Nobody’s on the Ball on AGI Alignment

AI Safety Fundamentals

00:00

Approaches to Achieving Alignment in AGI

The speaker discusses various approaches to achieving alignment in Artificial General Intelligence (AGI), including skepticism towards using math proofs, concerns about mechanistic interpretability, the use of reinforcement learning from human feedback, and a proposed plan for AGI alignment.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app