AI Safety Fundamentals: Alignment cover image

The Alignment Problem From a Deep Learning Perspective

AI Safety Fundamentals: Alignment

00:00

Internally Represented Goals in Deep Learning Models

This chapter explores the concept of internally represented goals in deep learning models, discussing both model-based and model-free policies. It argues that as AI architectures become more expressive and policies generalize outside their training distributions, internally-represented goals will become more extensive, including broadly scoped goals that apply to long-time frames and wide ranges of tasks.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app