"The Cognitive Revolution" | AI Builders, Researchers, and Live Player Analysis cover image

Dodging Latent Space Detectors: Obfuscated Activation Attacks with Luke, Erik, and Scott.

"The Cognitive Revolution" | AI Builders, Researchers, and Live Player Analysis

CHAPTER

Enhancing Model Robustness through Obfuscation

This chapter explores advanced techniques for improving model robustness, particularly focusing on obfuscated adversarial training (OAT). The discussion emphasizes the iterative nature of attacks and the importance of model representations in defending against adversarial manipulation. It highlights the complexities of understanding harmfulness in AI, the dynamics of attack budgets, and the sophisticated interplay between attacker strategies and defender responses in machine learning.

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner