AXRP - the AI X-risk Research Podcast cover image

24 - Superalignment with Jan Leike

AXRP - the AI X-risk Research Podcast

CHAPTER

The Role of Humans in AI Alignment

I would be excited to be replaced by AI. I think humans should always stay in the loop somehow. There's this quote from planning for AGI and beyond that says, it's possible that AGI capable enough to accelerate its own progress because major changes happen surprisingly quickly. And then it says we think a slower takeoff is easier to make safe. So one thing I wonder is like if we make this really smart or, you know, human level alignment researcher that we then like effectively 10x or 100x or something, does that end up playing into this like recursive self improvement loop? You can't have recursion without also improving your alignment a lot. That's just no way that that

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner