
"SolidGoldMagikarp (plus, prompt generation)"
LessWrong (Curated & Popular)
00:00
Optimizing a Prompt to Maximize a Target Token Frequency
The code is available at a link here in the post if you want more detail. This kind of prompt generation is only possible because token embedding space has a kind of semantic coherence. Semantically related tokens tend to be found close together. We discovered this by carrying out K-means clustering over the embedding space of the GPT token set and found many clusters that are surprisingly robust to random initialization of the centroids. It's tricky because unlike pixel values, the inputs to LLM's are discrete tokens. But with regularization and explicit coercion we can make it work.
Play episode from 08:21
Transcript


