AXRP - the AI X-risk Research Podcast cover image

11 - Attainable Utility and Power with Alex Turner

AXRP - the AI X-risk Research Podcast

CHAPTER

How to Preserve the Utility of a Reward Function, Right?

Ri: It seems like you've got to get the space of reward functions approximately right for randaly sampling them to end up being useful. The actually useful objective isn't going to be a function of the whole world's day and like the temperature and the pressure and whatever other statistics, but it's like chunking things nnot objects, and like featurization and such. And i think if, well, it's true that you could get some rather strange auxiliary golfs,. I think that just using the same forment that the primary reward isn't, should generally work pretty well.

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner