
Rohin Shah
TalkRL: The Reinforcement Learning Podcast
Learning From Human Feedback Is Better Than Reward Learning
Ronan Fergusson: I can point to this one as a something surprising to me. It's like towards ad decision theoretic model of assistance, or something like that. And then there's also co operative inverse reenforcement learning from chi. The idea with this paper was just to take te models that had already been proposed in these papers and explain why they were so nice. Is supposed to am other things that the field could be doing.
00:00
Transcript
Play full episode
Remember Everything You Learn from Podcasts
Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.