4min chapter

Towards Data Science cover image

110. Alex Turner - Will powerful AIs tend to seek power?

Towards Data Science

CHAPTER

Is There a Broader Sense of Optimization That Doesn't Involve Reward Functions?

A. A. Dowd: It's been my perception that one of the things a alignment generally has been missing is just a concrete arena in which to reason about these systems. And so i think once we start crossing that point, you might start to see some bad behaviour. Thing is that redirectible agents often have power of seaking sentents,. at least in the settings we look at. The reward function determines whether the opical polisies. are good or bad for an agent. But if you give a randomly generato reward function for pakman, and its like, well, well, pak man's just going to tend to die, right? L is not

00:00

Get the Snipd
podcast app

Unlock the knowledge in podcasts with the podcast player of the future.
App store bannerPlay store banner

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode

Save any
moment

Hear something you like? Tap your headphones to save it with AI-generated key takeaways

Share
& Export

Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode