TalkRL: The Reinforcement Learning Podcast

Sai Krishna Gottipati

Aug 1, 2022
Ask episode
Chapters
Transcript
Episode notes
1
Introduction
00:00 • 2min
2
The Multi Agentiral Human in the Lup Lurning and Some of Its Applications
01:46 • 2min
3
How Did You Get Into Computer Vision?
03:26 • 4min
4
Human in the Loop Learning
07:18 • 2min
5
Cogment Paper
09:27 • 5min
6
The Use of an I and R L in Safety Critical Situations?
14:49 • 2min
7
Cogment Versus
16:39 • 2min
8
Asymmetric Self Play for Automatic Gold Discovery in Robotic Manipulation
18:20 • 3min
9
Is It Better to Train a Student to Find a Cold Condition?
20:51 • 3min
10
Is There a Chicken and Egg Problem?
23:32 • 2min
11
Multi Teacher Approach to Self Play in Deep Reinforcement Learning
25:19 • 2min
12
Multi Teachers - Is There Any Alternative to Fix the Distribution?
26:59 • 2min
13
Using Multiple Teachers Is a Good Idea
28:37 • 3min
14
The Relationship Between Paird and the Pard System
31:12 • 2min
15
The Interaction of Continuous Coordination
32:55 • 2min
16
Hanabi Is a Very Challenging Card Game
35:00 • 2min
17
Can Agents Collaborate?
36:38 • 2min
18
Co-Operator Game
38:57 • 2min
19
Is There an Alternative to Population Base Training?
40:49 • 3min
20
How to Navigate the Synthetically Ible Chemical Space
43:36 • 2min
21
How to Find the Reactant in a Continuous Space?
45:42 • 4min
22
How Do You Represent the Chemicals?
49:52 • 2min
23
Are You Using Graphic Representation in Drug Discovery?
51:40 • 3min
24
A Case of Coming Back to Chess
54:40 • 2min
25
Is Mith Missing in the Design?
56:51 • 3min
26
Inductive Bias
59:29 • 2min
27
Is There a Failure of the Function Proxy?
01:01:20 • 2min
28
What Do You Think About Explainability in Chess?
01:03:16 • 2min
29
Scaling Laws - Is Scaling All You Need?
01:05:15 • 3min