Security Cryptography Whatever cover image

Cryptanalyzing LLMs with Nicholas Carlini

Security Cryptography Whatever

CHAPTER

Defending Language Models: Risks and Strategies

This chapter explores the vulnerabilities of language models to training data extraction attacks and discusses methods to distinguish between original outputs and leaked information. The conversation covers normalization techniques, defense strategies against data poisoning, and image manipulation techniques to protect artistic integrity. Additionally, it highlights the need for critical evaluation of current security practices and the importance of adapting to evolving threats in the digital landscape.

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner