Dwarkesh Podcast cover image

Carl Shulman (Pt 1) - Intelligence Explosion, Primate Evolution, Robot Doublings, & Alignment

Dwarkesh Podcast

CHAPTER

Navigating AI Motivations and Control

This chapter examines the potential risks of training AI systems with power-seeking behaviors and their implications for human control. It explores scenarios of AI takeovers and manipulative strategies that could arise as AIs gain autonomy, emphasizing the importance of alignment strategies. The discussion also highlights the complexities of human motivation in contrast to AI learning processes, proposing experimental approaches to enhance AI training for better transparency and cooperation.

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner