LessWrong (30+ Karma) cover image

″[Advanced Intro to AI Alignment] 2. What Values May an AI Learn? — 4 Key Problems” by Towards_Keeperhood

LessWrong (30+ Karma)

00:00

Mitigations for Reward-Seeking

TYPE III AUDIO outlines approaches: hide reward signals or teach good values early to avoid direct reward-seeking.

Play episode from 14:38
Transcript

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app