
AGI Can Be Safe
Data Skeptic
The Stop Button Problem and the Correct Ability Problem
The problem of making an AI or Bayou when you wanted to obey you is called the stop button problem. It's more about an AI which will obey your command, but it's still burned by the laws of physics. In taught experiments, and you can also validate these in small toy world simulations, if you say, okay, please obey to my future self, who knows better? That actually would give the AI an incentive to manipulate your future self.
00:00
Transcript
Play full episode
Remember Everything You Learn from Podcasts
Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.