4min chapter

80,000 Hours Podcast cover image

#151 – Ajeya Cotra on accidentally teaching AI models to deceive us

80,000 Hours Podcast

CHAPTER

Why AI Alignment Could Be Hard With Modern Deep Learning

I find it so hard to think about this sensibly and I could say something very stupid. It's just like dealing with something out of the ordinary way of things that I have any experience of dealing with as a human being in ordinary life. The journalist Ezra Klein has recently been using the analogy of casting spells to summon creatures through a portal. Another analogy I've heard is that of aliens. And another variant on the alien one relies on the fact that once you train a model with X capacities, you'll probably be able to run a very large number of copies of that model.

00:00

Get the Snipd
podcast app

Unlock the knowledge in podcasts with the podcast player of the future.
App store bannerPlay store banner

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode

Save any
moment

Hear something you like? Tap your headphones to save it with AI-generated key takeaways

Share
& Export

Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode