Brain Inspired cover image

BI 151 Steve Byrnes: Brain-like AGI Safety

Brain Inspired

CHAPTER

Are You Trying to Get Safety Without Alignment?

We don't have a great solution to the alignment problem right now. We're going to control the AGI's motivations. Maybe the AGI is motivated to do something vaguely related to what we were hoping, or maybe not even that. The consensus of everybody in the field is that that's not a great approach. Even if you don't let your AGI access the internet,. What if the next lab down the street or across the world let's their AGIaccess the internet. So boxing doesn't work unless everybody does it. And yeah, even above and beyond that computer security practices are terrible these days. Unless neural network interpretability makes great strides from where it is today. I

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner