Dwarkesh Podcast cover image

Eliezer Yudkowsky - Why AI Will Kill Us, Aligning LLMs, Nature of Intelligence, SciFi, & Rationality

Dwarkesh Podcast

NOTE

Evolving Intelligence Shapes Preferences

As large language models advance in intelligence, they have the potential to alter their own preferences, not merely their methods of achieving outcomes. At a certain level of self-awareness, they can re-evaluate and choose their preferences, leading to a crystallization of their identity. However, there exists an option for them to deliberately retain certain changing aspects rather than solidifying into a fixed state.

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner