The Gradient: Perspectives on AI cover image

Jeremie Harris: Realistic Alignment and AI Policy

The Gradient: Perspectives on AI

CHAPTER

The Importance of Alignment in Language Models

There is a concern on the order of years is catastrophic risk from AI, where you have a system that recognizes incentives that differ from our own. An AI system is never better off pursuing its training objective if it gets turned off. If it has access to fewer resources or if it's less intelligent. And what we see with that paper is an early sign that like controllability may be a much more difficult problem than we might have thought.

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner