Faster convergence for diffusion models

149 pointsposted a year ago
by vsroy

4 Comments

fxtentacle

a year ago

The title is not wrong, but it also doesn't feel correct either. What they do here is they use a pre-trained model to guide the training of a 2nd model. Of course, that massively speeds up training of the 2nd model. But it's not like you can now train a diffusion model from scratch 20x faster. Instead, this is a technique for transplanting an existing model onto a different architecture so that you don't have to start training from 0.

GaggiX

a year ago

I wonder how well this technique works if the distribution of the training dataset between the diffusion model and the image encoder is quite different, for example if you use DinoV2 as the encoder but train the diffusion model on anime.

gdiamos

a year ago

Still waiting for a competitive diffusion llm