Publication record · 18.cifr/2021.dhariwal.diffusion-classifier-guidance
18.cifr/2021.dhariwal.diffusion-classifier-guidanceWe show that diffusion models can achieve image sample quality superior to the current state-of-the-art generative models. We achieve this on unconditional image synthesis by finding a better architecture through a series of ablations. For conditional image synthesis, we further improve sample quality with classifier guidance: a simple, compute-efficient method for trading off diversity for fidelity using gradients from a classifier. We achieve an FID of 2.97 on ImageNet 128x128, 4.59 on ImageNet 256x256, and 7.72 on ImageNet 512x512.
Computing related research...
Loading DOI…
Sign in to run agents. GPU access requires an institutional membership.
How to get GPU access: Your university, lab, or company can become a CIFR institutional member. Members get GPU-accelerated runs for all their researchers. Contact us
No invocations yet — be the first to call this agent.
Classifier guidance requires a separate noisy-image classifier, adding complexity; classifier-free guidance and joint training are natural successors. Sampling speed (many forward passes) remains a bottleneck; distillation and improved samplers are flagged as open problems. Multi-modal and text-conditional extensions are obvious next steps.