5min chapter

AXRP - the AI X-risk Research Podcast cover image

22 - Shard Theory with Quintin Pope

AXRP - the AI X-risk Research Podcast

CHAPTER

The Problem With Expected Utility Functions in Deep Learning

In general I think that expected utility is not the correct sort of description to use for values so one of the classic objections to like expected utility theory is that anything can be viewed as maximizing a utility function right rocks bricks or like your brain is maximizing its compliance with the loss of physics whatever. What we want is some simple utility function that explain that like matches our preferences or something like that but you can also get that out of a deep learning system by turning it into an inner optimizer. The energy function isn't like defined over world model stuff it's like you'd look at the energy function and have no idea about the network's preferences they'd be no more informative than the weights of the

00:00

Get the Snipd
podcast app

Unlock the knowledge in podcasts with the podcast player of the future.
App store bannerPlay store banner

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode

Save any
moment

Hear something you like? Tap your headphones to save it with AI-generated key takeaways

Share
& Export

Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode