Hear This Idea

#66 – Michael Cohen on Input Tampering in Advanced RL Agents

Jun 25, 2023
Ask episode
Chapters
Transcript
Episode notes
1
Introduction
00:00 • 2min
2
The Problem With Gaussian Processes in Machine Learning
02:27 • 3min
3
The Difference Between Supervised and Unsupervised Learning
05:00 • 2min
4
The Differences Between Supervised and Unsupervised Learning
06:45 • 2min
5
How to Build an Image Generator Using Supervised Learning
08:56 • 2min
6
How to Use Chat GPT to Generate Images
10:56 • 3min
7
The Different Types of Reward Learning
13:50 • 3min
8
How to Train a Dog to Do Tricks
16:42 • 4min
9
How AI Intervenes in the Provision of Reward
20:17 • 2min
10
How to Give an AI a Model of the World
22:21 • 4min
11
How to Operationalize Reward for an Agent
26:25 • 4min
12
The Importance of Random Mutation in Reward Learning
30:07 • 3min
13
How to Be the Most Successful, Most Advanced RL Agent
32:40 • 2min
14
The Importance of Advanced Artificial Agents in Reinforcement Learning
34:27 • 2min
15
How to Expect an Agent to Understand the World
36:38 • 4min
16
The Distal Model and the Proximal Model of the World
41:04 • 4min
17
The Problem With Reward Modeling in AI
45:10 • 5min
18
The Cost of Experimentation Is Relatively Small
50:13 • 2min
19
How to Optimize the Proximal Model of an Artificial Agent
52:15 • 2min
20
The Cost of a Comment Defense System
54:18 • 2min
21
How to Hack Yourself and Not Take Over the World
55:59 • 2min
22
The Importance of Intervening in Computer Programming
57:50 • 3min
23
The Plausibility of Reward Maximizing Behavior
01:01:01 • 6min
24
Theoretical Arguments for Advanced AI
01:06:34 • 2min
25
The Limits of Advanced RL Agents
01:08:06 • 2min
26
The Multi-Agent Setting
01:09:57 • 5min
27
How to Create a Helper Agent
01:14:27 • 2min
28
The Multi-Agent Scenario
01:16:03 • 3min
29
The Instabilities of the World
01:18:40 • 3min
30
The Argument for Reward Is Not the Optimization Target
01:21:39 • 2min
31
The Alternative Framing of RL
01:23:28 • 3min
32
Reward Is Not the Optimization Target
01:26:53 • 3min
33
The Semantic Errors in Reward
01:29:53 • 3min
34
The Importance of Observation in a Chatbot
01:33:06 • 4min
35
The Difference Between Specification Gaming and Goal Misgeneralization
01:37:14 • 2min
36
The Limits of Goal Misgeneralization
01:39:10 • 2min
37
The Evolution of Inclusive Fitness
01:41:25 • 4min
38
Evolution and the Future of Genetic Fitness
01:45:29 • 3min
39
Evolution's Failure to Optimize Human Policies for Sperm Banks
01:48:12 • 2min
40
How to Avoid the Bad Outcome With Reinforcement Learning
01:49:54 • 2min
41
How to Combine Myopic Agents With Physical Isolation
01:52:15 • 4min
42
The Limitations of Boxing an Agent in AI
01:56:40 • 4min
43
The Myopia of Bazoo
02:00:27 • 2min
44
How to Make an Agent More Risk Averse
02:02:54 • 3min
45
The Power of Pessimistic Design for an Agent
02:05:56 • 2min
46
The Importance of Imitative Learning in AI
02:07:53 • 2min
47
The Different Models of Inverse Reinforcement Learning
02:10:18 • 3min
48
The Importance of Quantization in Reinforcement Learning
02:13:01 • 5min
49
The Importance of Uncertainty in Imitating Learning
02:18:19 • 3min
50
Rambo: A Practical Pessimistic Agent
02:20:50 • 3min
51
How to Make Safe Advanced AI
02:24:10 • 2min
52
How to Make Pessimism Disappear and Practice
02:26:19 • 2min
53
How to Raise a Child From Being an Expert in RL
02:28:33 • 3min