

Geometry-Aware Neural Rendering with Josh Tobin - #360
Mar 26, 2020
Josh Tobin, co-organizer of the Full Stack Deep Learning program and former research scientist at OpenAI, dives deep into geometry-aware neural rendering. He highlights the challenges in generating 3D scenes, the importance of domain randomization, and innovative methods bridging real-world data with simulations. The conversation also touches on the significance of encoder-decoder architectures in enhancing image rendering and how these techniques are revolutionizing AI applications in robotics.
AI Snips
Chapters
Transcript
Episode notes
Implicit Scene Understanding
- Implicit scene understanding helps robots act by creating a world representation from sensor observations.
- This contrasts with explicit representations, which become difficult in complex scenes.
Neural Rendering
- Neural rendering involves training a model to render a scene from any viewpoint, given observations from other viewpoints.
- Success implies an accurate implicit representation of the scene.
Geometry-Aware Neural Rendering
- Geometry-aware neural rendering builds on DeepMind's Generative Query Networks (GQN).
- It uses epipolar geometry to constrain the search for relevant pixels, improving rendering efficiency.