AI-powered
podcast player
Listen to all your favourite podcasts with AI-powered features
How to Make Diffusion Models Memorize a Small Fraction of the Data Set?
Stable diffusion is 900 million parameters. In practice, it's a couple gigabytes because you want a couple of bytes per weight. The data set that it was trained on though is enormous. It's like 10 terabytes. And so theoretically, it's just not going to be possible for the model to memorize all of the data. But in practice, what happens is it turns out that these models memorize not even a significant fraction of the data set.