
Special: Defeating AI Defenses (with Nicholas Carlini and Nathan Labenz)
Future of Life Institute Podcast
Navigating AI Vulnerabilities
This chapter explores the challenges of breaching AI defenses, particularly through adversarial examples in image classification models. It examines the interplay between attackers and defenders, emphasizing the attackers' adaptive advantages and strategies. Additionally, the discussion highlights the importance of simplicity in creating effective adversarial attacks and the implications for open-source AI safety.
00:00
Transcript
Play full episode
Remember Everything You Learn from Podcasts
Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.