
126 - Optimizing Continuous Prompts for Generation, with Lisa Li
NLP Highlights
Why Optimization Is Stable?
So just a couple of questions about why optimization might be unstable so right is it that when you already have a large network with the rest of the weights uh in a certain state you add this prefix drawn from say a random distribution of weights they're very different. Do you think that's why optimization is kind of unstable here do you have any intuitions on why that might be the case? umYeah definitely so i think it's probably one possible explanation like we didn't do off very deep into this question but intuitively it could be that because when we try to randomly initialize it lies in a very different space then then even if we are actually processing words and computing this latent activations however
00:00
Transcript
Play full episode
Remember Everything You Learn from Podcasts
Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.