The Jim Rutt Show

EP 325 Joe Edelman on Full-Stack AI Alignment

35 snips
Oct 7, 2025
Joe Edelman, a researcher focused on AI alignment, shares his insights on designing social systems that promote human flourishing. He discusses the importance of pluralism, critiquing conventional voting and market models for their superficiality. Edelman emphasizes the concept of 'thick models of value,' arguing that true values encompass deeper reasons and norms. He also addresses the risks of AI assistants potentially manipulating users and proposes innovative solutions like value-aware markets to navigate societal challenges. A thought-provoking conversation on the future of AI and governance!
Ask episode
AI Snips
Chapters
Books
Transcript
Episode notes
INSIGHT

Preferences Are Insufficient Signals

  • Preferencist models (clicks, votes) give shallow signals about people and miss why choices are made.
  • Joe Edelman argues alignment needs richer information like values and norms to guide AI and institutions.
INSIGHT

Language Alone Can Be Too Vague

  • Text-based specifications are expressive but often vague and underspecified for high-stakes norms.
  • Vague single-word goals like "helpful" or "harmless" leave open cultural interpretation and manipulation.
ADVICE

Be Precise When Specifying Model Behavior

  • Use philosophy and cognitive science to craft prompts that actually specify norms and values.
  • Measure how much detail is required rather than relying on vague words like "helpful."
Get the Snipd Podcast app to discover more snips from this episode
Get the app