LoGG tomorrow's paper: The Superposition of Diffusion Models Using the Itô Density Estimator
0 views
Skip to first unread message
Hannes Stärk
unread,
Feb 16, 2025, 5:47:13 PMFeb 16
Reply to author
Sign in to reply to author
Forward
Sign in to forward
Delete
You do not have permission to delete messages in this group
Copy link
Report message
Show original message
Either email addresses are anonymous for this group or you need the view member email addresses permission to view the original message
to lo...@googlegroups.com
Hi together,
Tomorrow's paper:
The Superposition of Diffusion Models Using the Itô Density Estimator https://arxiv.org/abs/2412.17762 (Marta Skreta,Lazar Atanackovic,Avishek Joey Bose,Alexander Tong,Kirill Neklyudov) The Cambrian explosion of easily accessible pre-trained diffusion models suggests a demand for methods that combine multiple different pre-trained diffusion models without incurring the significant computational burden of re-training a larger combined model. In this paper, we cast the problem of combining multiple pre-trained diffusion models at the generation stage under a novel proposed framework termed superposition. Theoretically, we derive superposition from rigorous first principles stemming from the celebrated continuity equation and design two novel algorithms tailor-made for combining diffusion models in SuperDiff. SuperDiff leverages a new scalable Itô density estimator for the log likelihood of the diffusion SDE which incurs no additional overhead compared to the well-known Hutchinson's estimator needed for divergence calculations. We demonstrate that SuperDiff is scalable to large pre-trained diffusion models as superposition is performed solely through composition during inference, and also enjoys painless implementation as it combines different pre-trained vector fields through an automated re-weighting scheme. Notably, we show that SuperDiff is efficient during inference time, and mimics traditional composition operators such as the logical OR and the logical AND. We empirically demonstrate the utility of using SuperDiff for generating more diverse images on CIFAR-10, more faithful prompt conditioned image editing using Stable Diffusion, and improved unconditional de novo structure design of proteins. this https URL
Speaker: Marta Skreta who is a fourth year Computer Science PhD student in the Matter Lab at the University of Toronto. Her research lies at the intersection of AI, chemistry, natural language processing, and self-driving labs.