Get the app
David Lindner
Research scientist on DeepMind's AGI Safety and Alignment team, focusing on myopic optimization and oversight methods such as MONA to mitigate multi-step reward hacking.
Best podcasts with David Lindner
Ranked by the Snipd community
Jun 15, 2025
• 1h 41min
43 - David Lindner on Myopic Optimization with Non-myopic Approval
chevron_right
David Lindner, a research scientist at DeepMind, dives into Myopic Optimization with Non-myopic Approval (MONA). He explains how MONA addresses multi-step reward hacking by ensuring actions align with human approval. The discussion covers MONA’s strengths and failure cases, comparing it to approaches like conservativism. Lindner highlights the importance of choosing approval signals carefully and the potential of MONA to create capable AI agents. The episode also explores practical deployment strategies and the complex dynamics of safety versus capability.
The AI-powered Podcast Player
Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
Get the app