Chapters
Transcript
Episode notes
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28
Introduction
00:00 • 2min
How to Deal With Human Bias?
02:01 • 4min
Reward Functions - Is That Really Imperative?
05:40 • 3min
Is There a Reasonable or a Simple Planning Module?
09:00 • 3min
Using a Neural Network to Make Optimum Predictions?
11:34 • 2min
Is Initialization a Good Approach?
13:27 • 2min
The Reward Functions With Ta
15:08 • 2min
Is the Utility of Lottery Tickets Really Realistic?
16:52 • 4min
Learning Models of the Human Demonstrator
21:21 • 3min
The Memory of Value in Iteration Networks Isn't Able to Express Literally Optimal Value
24:03 • 2min
I'm in Favor of Human Learning.
26:12 • 2min
The Difference Between Under Confident and Over Confident?
27:58 • 2min
How to Model Human Bias in a Logical Environment
30:23 • 3min
The Learning Planner Reward Function - How to Predict the Highest Reward
33:50 • 4min
Using a Differential Planner
37:46 • 2min
The Scientific Rigorousness of Machine Learning?
39:26 • 4min
Is It Possible to Learn a Reward Function From Human Behavior?
43:38 • 2min
Is There a Reward for Inferring What You Want?
45:55 • 2min
I Feel Like You Can Care About Your Life and Eat Cake?
48:17 • 2min
Generically Make Your Life Better?
50:00 • 3min
Is the City Task a Good Idea?
53:01 • 2min
Plan a City Rather Than Generically Make Your Life Better?
55:13 • 3min
Planning a City and Making Your Life Better
58:10 • 3min
Planning a City
01:00:43 • 2min
I Think It's a Good Idea to Have a Personal Assistant, or Something Like That.
01:02:20 • 2min
Is There a Core of Being Helpful?
01:03:58 • 2min
Ia, A, Is This a Domain Independent Corp?
01:05:32 • 2min
The Alignment News Letter
01:07:18 • 2min