AXRP - the AI X-risk Research Podcast cover image

24 - Superalignment with Jan Leike

AXRP - the AI X-risk Research Podcast

00:00

*2 aspects - Train and Validate AI Systems, for Deception testing *

The podcast discusses the importance of distinguishing between truly aligned AI systems and deceptively misaligned ones./nThe speaker mentions the need for validation methods to determine the version of the AI system./nValidation methods provide an independent signal to differentiate between the two versions of the AI systems.

Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner
Get the app