TalkRL: The Reinforcement Learning Podcast

Jordan Terry

Feb 22, 2022
Ask episode
Chapters
Transcript
Episode notes
1
Introduction
00:00 • 3min
2
Jim Ripo - What's Next?
03:13 • 5min
3
The Importance of Termination Versus Truncation
07:48 • 3min
4
The Difference Between Truncation and Photo Finish Ending?
10:32 • 2min
5
Is Discount Factor in Jin a Good Idea?
12:13 • 2min
6
Is Discounting an Environment or a Learning Co Property?
13:54 • 5min
7
Is There a Better Solution?
18:27 • 2min
8
The Problem Isn't Python, It's Python.
20:34 • 3min
9
The Limits of a Custom Agent
23:31 • 3min
10
Is There a Dependency on Tenser Flow?
26:30 • 2min
11
Is There a Difference Between Mujoko and Brack?
28:16 • 2min
12
The Jumpy Wrapper Road Map
30:40 • 3min
13
Petting Zoo Environment
33:11 • 2min
14
How to Multiplay Your Orel?
34:51 • 4min
15
Using a C Gm Api in a Game Model?
39:12 • 2min
16
Dos, How Does Reward Work in Pettings?
40:53 • 2min
17
Pettingso vs Pettingzoo?
42:46 • 2min
18
The Story of Petting Zoo Wrappers
44:53 • 2min
19
Is There a Scientifically Legitimate Experiment in R?
46:37 • 4min
20
The Impact of Implementation Specific Tricks on Black Box Optimizers
50:14 • 2min
21
Hyperopti or Optuna Hyper Perimeter Tuning?
52:32 • 3min
22
Is There a Benchmark in Malt Agent R L?
56:00 • 3min
23
What Is the Holy Grail of Your Line of Research?
59:19 • 2min
24
Is It a Good Idea to Insist on 50 Runs of Your Algar Inthem?
01:01:40 • 2min