"The Cognitive Revolution" | AI Builders, Researchers, and Live Player Analysis cover image

Dodging Latent Space Detectors: Obfuscated Activation Attacks with Luke, Erik, and Scott.

"The Cognitive Revolution" | AI Builders, Researchers, and Live Player Analysis

00:00

Enhancing Model Robustness through Obfuscation

This chapter explores advanced techniques for improving model robustness, particularly focusing on obfuscated adversarial training (OAT). The discussion emphasizes the iterative nature of attacks and the importance of model representations in defending against adversarial manipulation. It highlights the complexities of understanding harmfulness in AI, the dynamics of attack budgets, and the sophisticated interplay between attacker strategies and defender responses in machine learning.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app