UCL Discovery
UCL home » Library Services » Electronic resources » UCL Discovery

DEFT: Efficient Fine-tuning of Diffusion Models by Learning the Generalised h-transform

Denker, Alexander; Vargas, Francisco; Padhy, Shreyas; Didi, Kieran; Mathis, Simon V; Barbano, Riccardo; Dutordoir, Vincent; ... Lio, Pietro; + view all (2024) DEFT: Efficient Fine-tuning of Diffusion Models by Learning the Generalised h-transform. In: Proceedings of the Thirty-eighth Annual Conference on Neural Information Processing Systems. (pp. pp. 1-47). NeurIPS Green open access

[thumbnail of 11591_DEFT_Efficient_Fine_tuni-2.pdf]
Preview
Text
11591_DEFT_Efficient_Fine_tuni-2.pdf - Published Version

Download (13MB) | Preview

Abstract

Generative modelling paradigms based on denoising diffusion processes have emerged as a leading candidate for conditional sampling in inverse problems. In many real-world applications, we often have access to large, expensively trained unconditional diffusion models, which we aim to exploit for improving conditional sampling. Most recent approaches are motivated heuristically and lack a unifying framework, obscuring connections between them. Further, they often suffer from issues such as being very sensitive to hyperparameters, being expensive to train or needing access to weights hidden behind a closed API. In this work, we unify conditional training and sampling using the mathematically well-understood Doob's h-transform. This new perspective allows us to unify many existing methods under a common umbrella. Under this framework, we propose DEFT (Doob's h-transform Efficient FineTuning), a new approach for conditional generation that simply fine-tunes a very small network to quickly learn the conditional h-transform, while keeping the larger unconditional network unchanged. DEFT is much faster than existing baselines while achieving state-of-the-art performance across a variety of linear and non-linear benchmarks. On image reconstruction tasks, we achieve speedups of up to 1.6x, while having the best perceptual quality on natural images and reconstruction performance on medical images. Further, we also provide initial experiments on protein motif scaffolding and outperform reconstruction guidance methods.

Type: Proceedings paper
Title: DEFT: Efficient Fine-tuning of Diffusion Models by Learning the Generalised h-transform
Event: NeurIPS 2024
Location: Vancouver, Canada
Dates: 10th-15th December 2024
Open access status: An open access version is available from UCL Discovery
Publisher version: https://openreview.net/forum?id=AKBTFQhCjm
Language: English
Additional information: © The Authors 2024. Original content in this paper is licensed under the terms of the Creative Commons Attribution 4.0 International (CC BY 4.0) Licence (https://creativecommons.org/licenses/by/4.0/).
Keywords: Inverse problems, Generative Modelling, Diffusion Models, Conditional Generative Modelling, Diffusion model guidance
UCL classification: UCL
UCL > Provost and Vice Provost Offices > UCL BEAMS
UCL > Provost and Vice Provost Offices > UCL BEAMS > Faculty of Engineering Science > Dept of Computer Science
URI: https://discovery.ucl.ac.uk/id/eprint/10204141
Downloads since deposit
Loading...
17Downloads
Download activity - last month
Loading...
Download activity - last 12 months
Loading...
Downloads by country - last 12 months
Loading...

Archive Staff Only

View Item View Item