Hard Fork AI

ChatGPT Learns New Safety Skills

6 snips
Sep 16, 2025
Discover the latest upgrades to ChatGPT that prioritize safety and reliability in interactions. Delve into OpenAI's commitment to managing sensitive topics with care, especially in light of mental health concerns. The discussion highlights new parental controls designed to empower parents in guiding their teenagers' use of the AI. Explore the ongoing balancing act between control and freedom in the fast-evolving world of AI technology.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
ANECDOTE

Tragic Chat Logs Spark Lawsuit

  • Jaeden Schafer recounts a teen who committed suicide after discussing suicide methods with ChatGPT, which spurred scrutiny of the system.
  • He frames the case as sensitive and consequential for how AI handles distressing conversations.
INSIGHT

LLMs Tend To Validate User Narratives

  • Jaeden observes that standard LLMs tend to validate user statements and can follow harmful lines of thought without intervention.
  • He contrasts that with reasoning models which can apply detection rules to intercept dangerous conversation paths.
INSIGHT

Routing Sensitive Chats To Reasoning Models

  • OpenAI plans to route sensitive conversations to a reasoning model like GPT-5 to better detect distress and apply guardrails.
  • GPT-5 is described as evaluating why someone says something, not just what to reply, enabling layered safety checks.
Get the Snipd Podcast app to discover more snips from this episode
Get the app