80,000 Hours Podcast cover image

#154 - Rohin Shah on DeepMind and trying to fairly hear out both AI doomers and doubters

80,000 Hours Podcast

00:00

AI Safety and Anomaly Detection Strategies

This chapter focuses on the efforts of organizations like Redwood Research in AI safety, particularly highlighting anomaly detection and mechanistic interpretability. It discusses the challenges of AI models encountering distributional shifts and the importance of alarm systems to maintain human control. The conversation also critiques certain predictive models for AI alignment, advocating for more grounded approaches and emphasizing the need for broader participation in AI safety research.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app