Future of Life Institute Podcast cover image

Special: Defeating AI Defenses (with Nicholas Carlini and Nathan Labenz)

Future of Life Institute Podcast

CHAPTER

Navigating AI Vulnerabilities

This chapter explores the challenges of breaching AI defenses, particularly through adversarial examples in image classification models. It examines the interplay between attackers and defenders, emphasizing the attackers' adaptive advantages and strategies. Additionally, the discussion highlights the importance of simplicity in creating effective adversarial attacks and the implications for open-source AI safety.

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner