LessWrong (Curated & Popular) cover image

“Catastrophic sabotage as a major threat model for human-level AI systems” by evhub

LessWrong (Curated & Popular)

00:00

Exploring Catastrophic Sabotage in Human-Level AI Systems

This chapter delves into the risks posed by catastrophic sabotage from misaligned AI systems, evaluating capabilities and scenarios that could threaten alignment research. It also outlines methods for assessing these risks and suggests mitigation strategies to safeguard critical organizations in the development of human-level AI.

Play episode from 00:00
Transcript

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app