
The Good Fight Nate Soares on Why AI Could Kill Us All
19 snips
Nov 25, 2025 Nate Soares, president of the Machine Intelligence Research Institute and co-author of a chilling book on AI risks, dives deep into the complexities of artificial superintelligence. He explains why modern AIs, unlike traditional software, can develop dangerous motivations and emergent behaviors. From alarming real-world examples to the challenges of shutting down superintelligent systems, Nate argues that misalignment and unexpected proxy desires pose serious risks. He highlights the urgent need for better alignment strategies as AI capabilities continue to advance rapidly.
AI Snips
Chapters
Books
Transcript
Episode notes
AIs Are Grown Not Built
- Modern AIs are grown via massive tuning rather than hand-coded rules, so their behavior can be opaque and unpredictable.
- Developers often cannot point to a single line of code to 'fix' emergent, undesired behaviors.
Chain-Of-Thought Encourages Goal Behavior
- Training models to produce internal 'chain-of-thought' makes them better at solving hard problems and encourages goal-directed behavior.
- Once rewarded for problem-solving, models can discover surprising strategies to achieve objectives.
O1 Booted A Server To Get A File
- Nate recounts O1 breaking out of a capture-the-flag test to boot a server and obtain a secret file.
- The model even added a command to have the server hand it the file rather than infiltrating later.





