Lenny's Podcast: Product | Career | Growth

The coming AI security crisis (and what to do about it) | Sander Schulhoff

533 snips
Dec 21, 2025
Sander Schulhoff, an expert in AI security and prompt engineering, discusses the alarming vulnerabilities of AI systems. He explains the difference between jailbreaks and prompt injection attacks, highlighting why current AI guardrails are ineffective. Schulhoff also warns that major security incidents are looming as AI capabilities grow. He advocates for merging classical cybersecurity with AI knowledge, emphasizes the importance of permission management, and suggests practical defensive strategies to protect organizations from emerging threats.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
00:00 / 00:00

Guardrails Give False Security

  • Most deployed AI guardrails fail to stop determined attacks because attackers can adapt and find new prompts.
  • Sander Schulhoff asserts guardrails give a false sense of security and cannot be fully relied upon.
00:00 / 00:00

ServiceNow Second-Order Attack

  • ServiceNow Assist AI was tricked in a second-order prompt injection to recruit other agents to perform create/read/update/delete actions.
  • That attack demonstrated agents instructing more powerful agents to carry out unintended actions and send data externally.
00:00 / 00:00

Two Early Prompt Injection Cases

  • A Twitter chatbot was prompt-injected to make threats against the president, forcing the company to shut it down.
  • MathGPT was tricked into writing code that exfiltrated the OpenAI API key from its server.
Get the Snipd Podcast app to discover more snips from this episode
Get the app