
126 - Optimizing Continuous Prompts for Generation, with Lisa Li
NLP Highlights
00:00
Why Optimization Is Stable?
So just a couple of questions about why optimization might be unstable so right is it that when you already have a large network with the rest of the weights uh in a certain state you add this prefix drawn from say a random distribution of weights they're very different. Do you think that's why optimization is kind of unstable here do you have any intuitions on why that might be the case? umYeah definitely so i think it's probably one possible explanation like we didn't do off very deep into this question but intuitively it could be that because when we try to randomly initialize it lies in a very different space then then even if we are actually processing words and computing this latent activations however
Transcript
Play full episode