The Gradient: Perspectives on AI

Pete Florence: Dense Visual Representations, NeRFs, and LLMs for Robotics

Jan 5, 2023
Ask episode
Chapters
Transcript
Episode notes
1
Introduction
00:00 • 3min
2
How Many Quadcopters Did You Break?
02:45 • 5min
3
Using Diffuse Visual Representations of Objects in Computer Vision
08:06 • 4min
4
How Does a Diffuse 3D Modeling Work in Computer Vision?
11:45 • 3min
5
Can You Leverage Visual Motor Learning for Robotics?
14:34 • 4min
6
Robotic Manipulation
18:35 • 3min
7
NERF - What Is It All About?
21:48 • 4min
8
NERF for Pose Estimation of Objects
25:47 • 4min
9
How to Use NERF to Learn Dense Object Descriptors?
29:42 • 4min
10
Can We Just Generically Make Policy Learning Work Better?
33:51 • 5min
11
Socratic Models and Large Language Models
38:46 • 6min
12
Is Socratic Models Really a Viable Approach?
44:31 • 3min
13
How Does the Inner Model Log Work for Robotics?
47:36 • 5min
14
Using Language Models for Controlling Robots and Planning
52:11 • 5min
15
Code as Policies
57:22 • 2min
16
The Interaction of Language and Robotics in the Real World
59:05 • 5min
17
Robotics Is Hard, Right?
01:03:43 • 3min
18
The Limitations of Language Models in Robotics
01:06:56 • 4min
19
Is Language Models Really Helping With Vision and Action Fusion?
01:10:28 • 3min
20
Andre, Is There a Limitation in Robotics?
01:13:31 • 2min