
LessWrong (Curated & Popular) "Discussion with Nate Soares on a key alignment difficulty" by Holden Karnofsky
Apr 5, 2023
Chapters
Transcript
Episode notes
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
Introduction
00:00 • 2min
How an AI Avoids Pewter and POUDA
02:21 • 4min
Holden's Hypothetical AI Training Approach
06:49 • 2min
The Dangers of Needle Moving Alignment Research
09:13 • 2min
The Importance of Reflection in AI Research
11:18 • 4min
The Dangers of Mechanistic AI Training
15:24 • 3min
How AI's Can Help You Achieve Your Goals
18:51 • 4min
The Importance of Random Goals in AI Training
22:38 • 2min
The Implications of Holden's Training Setup
24:48 • 2min
The Importance of Robust Puda Avoidance
27:08 • 3min
The Game of Confidence
29:54 • 2min
The Doom of AI Training
31:25 • 2min
How Creative Intellectual Work Works
33:27 • 2min
The Future of Alignment Research
35:27 • 2min
AI-Driven Research on the Alignment Problem
37:51 • 2min
