Chapters
Transcript
Episode notes
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36
Introduction
00:00 • 2min
Reward Functions in Reaper Learning
01:55 • 5min
The Top Two Approaches
06:35 • 2min
Is It Possible to Learn the Concept of a Waterfall?
08:19 • 2min
Learning From Human Feedback
10:27 • 2min
Recommenders and Recommendation Systems
12:37 • 3min
Defisef - Is There a Robotic Control System?
15:23 • 4min
The Relationship Between Inverse Reenforcement Learning and Reward Learning
19:03 • 3min
I'll Say This - Sorry, Sorry, I Don't Think It Was at All
21:41 • 3min
Is the World Not How We Want It to Be?
24:36 • 2min
Learning From Human Feedback Is Better Than Reward Learning
27:03 • 4min
The Benefits of Assistive Robotics
30:49 • 2min
The Highest Level of Task Complexity in Deep Learning
33:13 • 2min
What's the Difference Between Assistance and Learning?
35:16 • 2min
Assistants Paradime - What Are the Benefits of Active Learning?
36:54 • 5min
The Reward Learning Paradime
42:20 • 2min
Inter Active Reward Learning
43:59 • 2min
Algorithms That Optimize Over Assistance
46:05 • 2min
The Next Paper on the Utility of Learning About Humans for Human Ai Coordination
47:49 • 4min
How to Get Robustness to Environments in Collaborative Games
51:46 • 3min
Using Humans in the Training Loop, Is That Really the Case?
54:39 • 3min
How to Evaluate the Robustness of Collaborative Agents
57:58 • 2min
Test Distributions
59:39 • 2min
The Three Types of Robustness in Reactor Learning Agents
01:01:12 • 2min
Scaling Up Deep Learning
01:02:59 • 3min
Is There a Canonical Definition of Air Linement?
01:05:59 • 3min
Ai Systems - How Does Alignment Relate to Ai Safety?
01:08:42 • 5min
What's Happening With Your Alignment News Letter?
01:14:03 • 3min
Alignment News Letter - I Highly Recommend It
01:16:40 • 5min
The Alignment Forum - Is That Right?
01:21:46 • 2min
Are There Any Alignment Issues in Science Fiction?
01:23:59 • 2min
How to Draw More Attention From the Academic Community?
01:25:55 • 3min
How to Approach the Alignment Problem When Faced With Heterogeneous Behaviors
01:28:30 • 3min
How Do We Best Handle Bias When Learning From Human Expert Demonstrations?
01:31:04 • 2min
The Holy Grail for Ai Systems Training
01:32:54 • 2min
Do You Have a Research Career Plan?
01:34:34 • 2min