Negative instructions may not work effectively in language models, as they can introduce the concept into context and make it difficult to negate. However, explicitly stating what not to do in detail can work well. Language models like GPT respond decently well to negative instructions. Moreover, having the model generate its own examples can be effective, as it can recall its own canonical examples and move on to solving the problem. Using self-generated examples and chain of thought rationales can be useful, but human-written examples may still be better for accuracy.

Get the Snipd
podcast app

Unlock the knowledge in podcasts with the podcast player of the future.
App store bannerPlay store banner

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode

Save any
moment

Hear something you like? Tap your headphones to save it with AI-generated key takeaways

Share
& Export

Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode