Get the app
Jeremy Gillen
Former MIRI research fellow and current AI safety researcher who is skeptical that aiming for corrigibility is a feasible short-term engineering strategy for preventing catastrophic AI outcomes.
Best podcasts with Jeremy Gillen
Ranked by the Snipd community
13 snips
Nov 14, 2025
• 2h 18min
The AI Corrigibility Debate: MIRI Researchers Max Harms vs. Jeremy Gillen
chevron_right
Max Harms, an AI alignment researcher and author of the novel Red Heart, debates with former MIRI research fellow Jeremy Gillen on AI corrigibility. Max argues that aiming for obedient, corrigible AI is essential to prevent existential risks, drawing parallels to human assistant dynamics. Jeremy is skeptical about the feasibility of this approach as a short-term solution. The discussion explores the intricacies of maintaining control over superintelligent AI and whether efforts toward corrigibility might be a hopeful strategy or an over-optimistic dream.
The AI-powered Podcast Player
Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
Get the app