4min chapter

The Inside View cover image

Shahar Avin–Intelligence Rising, AI Governance

The Inside View

CHAPTER

How to Red Team AI Systems

Red teaming is an adversarial relationship where you pretend to be the adversary and break whatever system or process you have in place. I think there's more and more adversarial examples when building self-driving cars. With the rising of language models, people can just test models like HP3 and find ways in which to make these behaviors. You might worry that when you get to very advanced systems, this method no longer works and you need different kinds of solutions.

00:00

Get the Snipd
podcast app

Unlock the knowledge in podcasts with the podcast player of the future.
App store bannerPlay store banner

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode

Save any
moment

Hear something you like? Tap your headphones to save it with AI-generated key takeaways

Share
& Export

Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode