The 80000 Hours Podcast on Artificial Intelligence cover image

Two: Ajeya Cotra on accidentally teaching AI models to deceive us

The 80000 Hours Podcast on Artificial Intelligence

NOTE

Understanding AI Alignment through Analogies

Reasoning about AI alignment is challenging as it involves dealing with concepts outside of human experience. Analogies like predicting octopus behavior based on sea snail ancestors highlight the difficulty in understanding and predicting AI behavior, resulting in diverse perspectives among experts.

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner