LessWrong (Curated & Popular) cover image

"SolidGoldMagikarp (plus, prompt generation)"

LessWrong (Curated & Popular)

00:00

Optimizing a Prompt to Maximize a Target Token Frequency

The code is available at a link here in the post if you want more detail. This kind of prompt generation is only possible because token embedding space has a kind of semantic coherence. Semantically related tokens tend to be found close together. We discovered this by carrying out K-means clustering over the embedding space of the GPT token set and found many clusters that are surprisingly robust to random initialization of the centroids. It's tricky because unlike pixel values, the inputs to LLM's are discrete tokens. But with regularization and explicit coercion we can make it work.

Play episode from 08:21
Transcript

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app