A New Trick Uses AI to Jailbreak AI Models—Including GPT-4
Dec 11, 2023
05:28
forum Ask episode
view_agenda Chapters
auto_awesome Transcript
info_circle Episode notes
Adversarial algorithms can systematically probe large language models like OpenAI’s GPT-4 for weaknesses that can make them misbehave. Read the story here.