The Gradient: Perspectives on AI cover image

Jeremie Harris: Realistic Alignment and AI Policy

The Gradient: Perspectives on AI

CHAPTER

How to Maximize Your Debate

The key to me is like, I don't expect the system to care about anything other than its objective. A priority makes it very difficult to claim that we could expect this thing to learn to care about whatever we tried to get it to learned to care about. The idea of orthogonality can be totally de correlated with intelligence and goals. It's not if it understands that you want it to make paper clips. Great. That doesn't mean it's going to ignore its implicit incentives to seek power just because they are in an adversarial relationship relative to humans.

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner