
Eliezer Yudkowsky - Why AI Will Kill Us, Aligning LLMs, Nature of Intelligence, SciFi, & Rationality
Dwarkesh Podcast
Evolving Intelligence Shapes Preferences
As large language models advance in intelligence, they have the potential to alter their own preferences, not merely their methods of achieving outcomes. At a certain level of self-awareness, they can re-evaluate and choose their preferences, leading to a crystallization of their identity. However, there exists an option for them to deliberately retain certain changing aspects rather than solidifying into a fixed state.
00:00
Transcript
Play full episode
Remember Everything You Learn from Podcasts
Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.