Future of Life Institute Podcast cover image

Connor Leahy on the State of AI and Alignment Research

Future of Life Institute Podcast

CHAPTER

The Core Problem of Alignment in RLETF

The core problem of alignment is how do you get a very complex powerful system that you don't understand to reliably do something complicated in domains where you cannot supervise it? RLETF does not address this problem. It doesn't even claim to suggest this problem. There's no reason to expect that RLETF should solve this problem. This is like, it's like, you know, clicker training in a alien.

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner