The 80000 Hours Podcast on Artificial Intelligence cover image

Two: Ajeya Cotra on accidentally teaching AI models to deceive us

The 80000 Hours Podcast on Artificial Intelligence

CHAPTER

Analyzing the Evaluation Process of AI Plans and Human Feedback in Reinforcement Learning

Exploring the comparison between outcomes-based and plan-making AI systems, emphasizing on the role of human feedback and reward models in training AI systems.

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner