The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence)

Watermarking Large Language Models to Fight Plagiarism with Tom Goldstein - 621

Mar 20, 2023
In this discussion, Tom Goldstein, an associate professor at the University of Maryland specializing in AI security and safety, dives into his pioneering research on watermarking large language models. He explains how these watermarks help combat misinformation and the mechanisms behind tracking AI-generated content. Tom also examines the economic and ethical implications of watermarking, blending profit with social responsibility. Additionally, he touches on challenges with data leakage in diffusion models and scaling plagiarism detection across massive datasets.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
ANECDOTE

Invisibility Cloak Project

  • Tom Goldstein's invisibility cloak project aimed to create clothing that evades object detectors.
  • The project involved printing paper mock-ups and testing various prototypes.
INSIGHT

Model Sensitivity of Adversarial Patterns

  • Adversarial patterns' effectiveness varies significantly across object detection models.
  • A pattern trained on one model might not transfer well to another, making reliable evasion difficult.
INSIGHT

How LLM Watermarking Works

  • LLMs generate text word by word, using a vocabulary of possible words.
  • Watermarking involves pseudo-randomly partitioning this vocabulary into "green" (good) and "red" (bad) lists.
Get the Snipd Podcast app to discover more snips from this episode
Get the app