
Ep 58: Sam Bowman on ChatGPT & Controlling AI
Brave New World -- hosted by Vasant Dhar
Alignment Problems
Alignment is the problem of taking that system and making it do the thing that you want, sort of giving it a goal and having it pursue that goal. And it can be hard. All of the sort of silly, silly failure demos that you see on Twitter with chat2PT are sort of classic examples of alignment failure. You've got problems where models don't quite learn the goals that we try to give them or even if they do learn the goals doesn't always work out as planned.
00:00
Transcript
Play full episode
Remember Everything You Learn from Podcasts
Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.