
Cryptanalyzing LLMs with Nicholas Carlini
Security Cryptography Whatever
Defending Language Models: Risks and Strategies
This chapter explores the vulnerabilities of language models to training data extraction attacks and discusses methods to distinguish between original outputs and leaked information. The conversation covers normalization techniques, defense strategies against data poisoning, and image manipulation techniques to protect artistic integrity. Additionally, it highlights the need for critical evaluation of current security practices and the importance of adapting to evolving threats in the digital landscape.
00:00
Transcript
Play full episode
Remember Everything You Learn from Podcasts
Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.