4min chapter

AXRP - the AI X-risk Research Podcast cover image

24 - Superalignment with Jan Leike

AXRP - the AI X-risk Research Podcast

CHAPTER

Neural Networks Generalize Across Languages

The hope is that maybe interpretability could tell you things about like it's got some kernel of lying in there but it only unlocks here. I think fundamentally this is also a really interesting machine learning problem just like how do you how do neural networks actually generalize outside of the IID setting and what are the mechanisms that work here? We don't know why or we don't have an answer to this question so far no-one does but it seems like a thing that's really important to understand yeahYeah, with causal modeling you're gonna need some sort of like causal modeling there from alleged beliefs to outputs right?" "You might you might need a lot more," he says. '

00:00

Get the Snipd
podcast app

Unlock the knowledge in podcasts with the podcast player of the future.
App store bannerPlay store banner

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode

Save any
moment

Hear something you like? Tap your headphones to save it with AI-generated key takeaways

Share
& Export

Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode