Chapters
Transcript
Episode notes
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46
Introduction
00:00 • 2min
The Importance of Human-Level Automated Alignment Research
02:09 • 3min
How Important Is the Human Level Qualifier in Alignment Research?
05:31 • 4min
How to Scale Up an Automated Alignment Research Model
09:21 • 3min
How to Make an Automated Alignment Researcher
12:22 • 2min
Automating 99.9% of Alignment Research
14:19 • 2min
The Alignment Tax
16:14 • 2min
Scalable Oversight for AI Alignment Research
18:44 • 2min
How to Align Super Intelligence
20:40 • 2min
The Role of Humans in AI Alignment
22:58 • 3min
How to Make a Smart AI Alignment Researcher
26:16 • 4min
How to Scalable Oversight With Alignment Research
30:10 • 2min
The Importance of Using a Criticism Model in AI Alignment Research
32:36 • 2min
The Discriminator Criticism App
34:53 • 5min
Scale Below Recid - How to Measure Scale Below Recid
39:26 • 2min
The Problems With Automated Task Evaluation
41:17 • 2min
The Future of Interpretability
43:09 • 2min
The Importance of Interpretability in Language Models
45:26 • 4min
Automated Interpretability for Neurons
49:56 • 4min
The Importance of Scaling Interpretability
53:46 • 2min
The Risks of Misalignment of AI Systems
55:19 • 2min
How to Train Misaligned Models to Be Consistent Liars
57:42 • 2min
How to Train a System to Succeed
01:00:02 • 3min
The Core Technical Challenges of Super Intelligence Alignment
01:02:36 • 2min
The Four Years of AI Progress
01:04:29 • 3min
The Alignment Problem
01:07:06 • 3min
The Importance of Good Measures in Audits of AI Systems
01:10:16 • 3min
How the Open AI Team Is Relating to the Alignment Team
01:13:44 • 3min
How the Super Alignment Team Is Relating to Other Things at Open EI Like Efforts to Make Chat GPT Nicer Minimize on Our Sources
01:16:45 • 2min
The Importance of Collaboration in AI Research
01:18:18 • 2min
The Advantages of Automated Alignment Research
01:20:46 • 4min
OpenEI's Plan for AI Alignment
01:24:22 • 2min
How to Scale a New Superignment Team
01:26:12 • 2min
Generalization and Scalable Oversight
01:28:10 • 3min
The Importance of Generalization in Neural Networks
01:31:08 • 4min
The Interaction of Cross Validation and Interpretation
01:34:40 • 2min
The Importance of Cross Validation Techniques
01:36:31 • 2min
Neural Networks Generalize Across Languages
01:38:23 • 4min
How to Summon the Complexity Theoretic Definition of Love and Goodness Within the Super Alignment Team
01:42:03 • 2min
How to Make AI Systems More Aligned
01:43:45 • 5min
The Future of Alignment
01:48:30 • 5min
The Importance of Scalable Oversight
01:53:45 • 5min
The Importance of Language Models for Alignment
01:58:32 • 3min
How to Improve Pre-Training Loss
02:01:33 • 1min
The Benefits of Language Models
02:02:56 • 2min
How to Align Super Intelligence in Four Years
02:05:05 • 3min