

Ajeya Cotra on how Artificial Intelligence Could Cause Catastrophe
8 snips Nov 3, 2022
Chapters
Transcript
Episode notes
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20
Introduction
00:00 • 2min
AI Safety Research: A Superset of AI Alignment Research
02:05 • 5min
How AI Can Be Capable?
06:40 • 3min
How to Train a Model to Predict the World
09:10 • 3min
The Importance of General Planning in Language Models
11:52 • 3min
How to Reduce the Failure Rate of Language Models
14:44 • 3min
The Future of AI
18:08 • 2min
The Importance of Self-Reflection in Human Psychology
19:48 • 2min
The Naive Safety Effort Hypothesis
22:03 • 2min
The Role of Alex in the Development of Goals
24:07 • 3min
The Future of Deception
26:37 • 2min
How to Train Your Robot to Be Smart and Deceptive
29:04 • 2min
The Importance of Human Feedback in Decision-Making
31:08 • 3min
The Prototypical Action of Alexa
34:38 • 2min
The Future of Alexa
36:24 • 2min
The Self-Preserving Instinct of Alex
38:39 • 5min
How to Train Alex to Be Honest
43:18 • 4min
Inverse Reinforcement Learning: A Possible Way Forward for Alex
47:10 • 3min
Inverse Reinforcement Learning on Humans in Their Best State
49:41 • 2min
The Future of Interpretability
51:22 • 3min