AXRP - the AI X-risk Research Podcast

2 - Learning Human Biases with Rohin Shah

Dec 11, 2020
Ask episode
Chapters
Transcript
Episode notes
1
Introduction
00:00 • 2min
2
How to Deal With Human Bias?
02:01 • 4min
3
Reward Functions - Is That Really Imperative?
05:40 • 3min
4
Is There a Reasonable or a Simple Planning Module?
09:00 • 3min
5
Using a Neural Network to Make Optimum Predictions?
11:34 • 2min
6
Is Initialization a Good Approach?
13:27 • 2min
7
The Reward Functions With Ta
15:08 • 2min
8
Is the Utility of Lottery Tickets Really Realistic?
16:52 • 4min
9
Learning Models of the Human Demonstrator
21:21 • 3min
10
The Memory of Value in Iteration Networks Isn't Able to Express Literally Optimal Value
24:03 • 2min
11
I'm in Favor of Human Learning.
26:12 • 2min
12
The Difference Between Under Confident and Over Confident?
27:58 • 2min
13
How to Model Human Bias in a Logical Environment
30:23 • 3min
14
The Learning Planner Reward Function - How to Predict the Highest Reward
33:50 • 4min
15
Using a Differential Planner
37:46 • 2min
16
The Scientific Rigorousness of Machine Learning?
39:26 • 4min
17
Is It Possible to Learn a Reward Function From Human Behavior?
43:38 • 2min
18
Is There a Reward for Inferring What You Want?
45:55 • 2min
19
I Feel Like You Can Care About Your Life and Eat Cake?
48:17 • 2min
20
Generically Make Your Life Better?
50:00 • 3min
21
Is the City Task a Good Idea?
53:01 • 2min
22
Plan a City Rather Than Generically Make Your Life Better?
55:13 • 3min
23
Planning a City and Making Your Life Better
58:10 • 3min
24
Planning a City
01:00:43 • 2min
25
I Think It's a Good Idea to Have a Personal Assistant, or Something Like That.
01:02:20 • 2min
26
Is There a Core of Being Helpful?
01:03:58 • 2min
27
Ia, A, Is This a Domain Independent Corp?
01:05:32 • 2min
28
The Alignment News Letter
01:07:18 • 2min