5min chapter

AXRP - the AI X-risk Research Podcast cover image

6 - Debate and Imitative Generalization with Beth Barnes

AXRP - the AI X-risk Research Podcast

CHAPTER

How to Interpret a Big Nural Net in a Test Set?

The problem of how to train a model for interpretability. How do you represent things in such a way that the human has meaningful understanding, and it's like reasonably efficient? And this, i think, ends up being pretty close to the sort of hard problems of interpretability. I mean, it seems like there are a lot of long treatises that you have to deal with. But we also, like, if you imagine representing everything that alphafold knows in text, it's just going to be horribly inefficient to try an like otise that and get the do it. The one that's like just trust this big black box, which seems sort of doable, but it

00:00

Get the Snipd
podcast app

Unlock the knowledge in podcasts with the podcast player of the future.
App store bannerPlay store banner

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode

Save any
moment

Hear something you like? Tap your headphones to save it with AI-generated key takeaways

Share
& Export

Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode