AXRP - the AI X-risk Research Podcast

1 - Adversarial Policies with Adam Gleave

Dec 11, 2020
Ask episode
Chapters
Transcript
Episode notes
1
Introduction
00:00 • 3min
2
How Much Training Did You Do to Train These Adversarial Policies?
02:35 • 2min
3
How to Train a Kicker to Block a Bull
04:13 • 2min
4
Rarbotic - A Multiagent Environment
06:05 • 2min
5
Observation Rate Increases With Self Placement Techniques
08:19 • 2min
6
Is It Like a Rule of Fum in a Training Setting?
10:44 • 2min
7
Is There a Difference Between the Input and the Output of a Robotic Environment?
13:07 • 2min
8
Is the Adversarial Victim Always the Kicker?
15:19 • 2min
9
What Happens When You Blindfold or Mask Victim Policies?
17:22 • 2min
10
Sumo Ant Sumo Wrestling
19:43 • 2min
11
How to Control a Victin Policy in Sumo?
21:22 • 2min
12
A, I Think We Didn't Choose a Particular Layer of a Network.
23:44 • 2min
13
Is the Space of Policies Transifret?
25:34 • 4min
14
Training Adversarial Policies
29:29 • 4min
15
A, You're Gonna Be Able to Discover a Blind Spot
33:44 • 2min
16
Kick and Defent
35:48 • 2min
17
Avas Writer, What's the Reception of the Work?
37:53 • 3min
18
Are There Any Misconceptions About Security?
40:32 • 2min
19
Is There a Preprint of This Paper?
42:33 • 4min
20
Is It Safe to Say That You're Interested in Ai Alignment?
46:06 • 2min
21
Is There a Problem in the Safety Community?
47:42 • 5min
22
Is There a Reliability of the Feedback Mechanism?
52:47 • 3min
23
Is There a Human Level Classification Accuracy on Natural Images?
55:30 • 3min