AI Safety Fundamentals: Alignment cover image

Can We Scale Human Feedback for Complex AI Tasks?

AI Safety Fundamentals: Alignment

00:00

Enhancing Feedback with Recursive Reward Modeling and Constitutional AI

Exploring the use of Recursive Reward Modeling and Constitutional AI to improve feedback quality in AI systems. Discussing how leveraging average humans and AI assistants can lead to better feedback for training AI models, along with the concept of using AI systems for generating feedback guidelines.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app