Astral Codex Ten Podcast cover image

Contra The xAI Alignment Plan

Astral Codex Ten Podcast

00:00

The Future of AI Alignment

I predicted if we ever got a maximally curious superintelligence, it would scan all humans. This doesn't mean AI can't have a goal of understanding the universe. It just can't be the whole alignment strategy. But also, we couldn't make a maximally curiosity AI even if we wanted to. We direct AI's through reinforcement learning, telling them to do certain things and avoid other things. The existing field of AI alignment tries to figure out how to install any goal at all into an AI.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app