Brave New World -- hosted by Vasant Dhar cover image

Ep 58: Sam Bowman on ChatGPT & Controlling AI

Brave New World -- hosted by Vasant Dhar

CHAPTER

Alignment Problems

Alignment is the problem of taking that system and making it do the thing that you want, sort of giving it a goal and having it pursue that goal. And it can be hard. All of the sort of silly, silly failure demos that you see on Twitter with chat2PT are sort of classic examples of alignment failure. You've got problems where models don't quite learn the goals that we try to give them or even if they do learn the goals doesn't always work out as planned.

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner