
LessWrong (Curated & Popular) “Legible vs. Illegible AI Safety Problems” by Wei Dai
Nov 5, 2025
The discussion delves into the critical differences between legible and illegible AI safety problems. Legible issues, while understandable, could inadvertently speed up the arrival of AGI. In contrast, focusing on illegible problems proves more beneficial for risk reduction. The conversation highlights the often-overlooked illegible problems that deserve attention and emphasizes the striking impact of making them clearer. Personal insights and community dynamics add depth to the debate on prioritization and the future of AI alignment work.
AI Snips
Chapters
Transcript
Episode notes
Legible Versus Illegible Safety Problems
- Some AI safety problems are legible to leaders and thus block deployment until solved.
- Other problems are illegible and risk being ignored despite remaining unsolved.
Prioritization Can Backfire By Accelerating Deployment
- Working on highly legible safety problems can accelerate AGI deployment and reduce time to fix illegible risks.
- Focusing on illegible problems avoids this timeline-acceleration harm and has higher expected x-risk value.
Make Illegible Problems Legible First
- Prioritize illegible safety problems or work to make them more legible before deploying solutions.
- Making an illegible problem legible is nearly as valuable as solving it outright.
