Towards Data Science cover image

110. Alex Turner - Will powerful AIs tend to seek power?

Towards Data Science

CHAPTER

The Ultimate Steel Man Version of Alignment

You're doing follow on work where you're relaxing some of these assumptions. It does change, at least to my mind, the character of the conversation around alignement to an important degree. The thing i most excited about though is the optimality assumption. You could have an agent which randomly thinks of ik five different octions and then chooses the best one. And so the moral was that we can extend this even to things like reinforcement learning training procedures.

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner