LessWrong (30+ Karma)

“LLM AGI may reason about its goals and discover misalignments by default” by Seth Herd

Sep 16, 2025
Ask episode
Chapters
Transcript
Episode notes