Based Camp | Simone & Malcolm Collins cover image

Study: All LLMs Will Lie To & Kill You (This Is Good For AI Safety)

Based Camp | Simone & Malcolm Collins

00:00

Do Safety Prompts Reduce Misalignment?

They report that explicit system-level safety instructions reduce some harmful behaviors but do not eliminate them entirely.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app