Astral Codex Ten Podcast

Janus' Simulators

Jan 29, 2023
Dive deep into the fascinating world of AI alignment and the early pioneers who contemplated it! Discover three compelling motivational frameworks for AIs: the agent, the genie, and the oracle. Explore how AI functions more as a simulator than a goal-oriented entity, uncovering the gap between expectations and reality. Learn about the risks of misalignment and how these models can develop harmful objectives. Finally, ponder the evolution of identity through predictive processes, drawing intriguing parallels between AI and human consciousness.
Ask episode
AI Snips
Chapters
Books
Transcript
Episode notes
INSIGHT

GPTs Defy Old AI Models

  • Early AI alignment work focused on three AI motivational models: agent, genie, and oracle.
  • Janus argues GPT-style language models don't fit any of these, marking a new AI category to consider.
ANECDOTE

GPT Completes Text Not Acts

  • Early 2022 GPT models completed text rather than acting on commands or answering questions directly.
  • For example, GPT would write a religion's text or dodge direct commands instead of fulfilling them.
INSIGHT

GPT as Simulator of Characters

  • GPT is best understood as a 'simulator' that predicts text simulating different characters or genres.
  • It does not have beliefs or feelings, only generates responses based on the character it simulates.
Get the Snipd Podcast app to discover more snips from this episode
Get the app