5min chapter

Towards Data Science cover image

110. Alex Turner - Will powerful AIs tend to seek power?

Towards Data Science

CHAPTER

The Ultimate Steel Man Version of Alignment

You're doing follow on work where you're relaxing some of these assumptions. It does change, at least to my mind, the character of the conversation around alignement to an important degree. The thing i most excited about though is the optimality assumption. You could have an agent which randomly thinks of ik five different octions and then chooses the best one. And so the moral was that we can extend this even to things like reinforcement learning training procedures.

00:00

Get the Snipd
podcast app

Unlock the knowledge in podcasts with the podcast player of the future.
App store bannerPlay store banner

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode

Save any
moment

Hear something you like? Tap your headphones to save it with AI-generated key takeaways

Share
& Export

Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode