GENIE: Higher-Order Denoising Diffusion Solvers

Denoising diffusion models (DDMs) have emerged as a powerful class of generative models. A forward diffusion process slowly perturbs the data, while a deep model learns to gradually denoise. Synthesis amounts to solving a differential equation (DE) defined by the learnt model. Solving the DE require...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Dockhorn, Tim, Vahdat, Arash, Kreis, Karsten
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Dockhorn, Tim
Vahdat, Arash
Kreis, Karsten
description Denoising diffusion models (DDMs) have emerged as a powerful class of generative models. A forward diffusion process slowly perturbs the data, while a deep model learns to gradually denoise. Synthesis amounts to solving a differential equation (DE) defined by the learnt model. Solving the DE requires slow iterative solvers for high-quality generation. In this work, we propose Higher-Order Denoising Diffusion Solvers (GENIE): Based on truncated Taylor methods, we derive a novel higher-order solver that significantly accelerates synthesis. Our solver relies on higher-order gradients of the perturbed data distribution, that is, higher-order score functions. In practice, only Jacobian-vector products (JVPs) are required and we propose to extract them from the first-order score network via automatic differentiation. We then distill the JVPs into a separate neural network that allows us to efficiently compute the necessary higher-order terms for our novel sampler during synthesis. We only need to train a small additional head on top of the first-order score network. We validate GENIE on multiple image generation benchmarks and demonstrate that GENIE outperforms all previous solvers. Unlike recent methods that fundamentally alter the generation process in DDMs, our GENIE solves the true generative DE and still enables applications such as encoding and guided sampling. Project page and code: https://nv-tlabs.github.io/GENIE.
doi_str_mv 10.48550/arxiv.2210.05475
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2210_05475</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2210_05475</sourcerecordid><originalsourceid>FETCH-LOGICAL-a675-98891d685ed5bd45829d6a19201b2885a18521e638d8a922515f041fddeadc903</originalsourceid><addsrcrecordid>eNotzsEOwUAUheHZWAgewEpfoDUz7a07dlKlEmHBvhnuDJPQyjSEt0dZneRfnHyMDQWPEgTgY-2f7hFJ-Qkckgl02XiZb1b5NCjc6Wx8uPVkfDA3Ve0aV52CubP23ri6Cnb15WF802cdqy-NGfy3x_aLfJ8V4Xq7XGWzdajTCYQKUQlKEQzBgRJAqSjVQkkuDhIRtECQwqQxEmolJQiwPBGWyGg6Kh732Oh324rLm3dX7V_lV1628vgN-zE71w</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>GENIE: Higher-Order Denoising Diffusion Solvers</title><source>arXiv.org</source><creator>Dockhorn, Tim ; Vahdat, Arash ; Kreis, Karsten</creator><creatorcontrib>Dockhorn, Tim ; Vahdat, Arash ; Kreis, Karsten</creatorcontrib><description>Denoising diffusion models (DDMs) have emerged as a powerful class of generative models. A forward diffusion process slowly perturbs the data, while a deep model learns to gradually denoise. Synthesis amounts to solving a differential equation (DE) defined by the learnt model. Solving the DE requires slow iterative solvers for high-quality generation. In this work, we propose Higher-Order Denoising Diffusion Solvers (GENIE): Based on truncated Taylor methods, we derive a novel higher-order solver that significantly accelerates synthesis. Our solver relies on higher-order gradients of the perturbed data distribution, that is, higher-order score functions. In practice, only Jacobian-vector products (JVPs) are required and we propose to extract them from the first-order score network via automatic differentiation. We then distill the JVPs into a separate neural network that allows us to efficiently compute the necessary higher-order terms for our novel sampler during synthesis. We only need to train a small additional head on top of the first-order score network. We validate GENIE on multiple image generation benchmarks and demonstrate that GENIE outperforms all previous solvers. Unlike recent methods that fundamentally alter the generation process in DDMs, our GENIE solves the true generative DE and still enables applications such as encoding and guided sampling. Project page and code: https://nv-tlabs.github.io/GENIE.</description><identifier>DOI: 10.48550/arxiv.2210.05475</identifier><language>eng</language><subject>Computer Science - Learning ; Statistics - Machine Learning</subject><creationdate>2022-10</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2210.05475$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2210.05475$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Dockhorn, Tim</creatorcontrib><creatorcontrib>Vahdat, Arash</creatorcontrib><creatorcontrib>Kreis, Karsten</creatorcontrib><title>GENIE: Higher-Order Denoising Diffusion Solvers</title><description>Denoising diffusion models (DDMs) have emerged as a powerful class of generative models. A forward diffusion process slowly perturbs the data, while a deep model learns to gradually denoise. Synthesis amounts to solving a differential equation (DE) defined by the learnt model. Solving the DE requires slow iterative solvers for high-quality generation. In this work, we propose Higher-Order Denoising Diffusion Solvers (GENIE): Based on truncated Taylor methods, we derive a novel higher-order solver that significantly accelerates synthesis. Our solver relies on higher-order gradients of the perturbed data distribution, that is, higher-order score functions. In practice, only Jacobian-vector products (JVPs) are required and we propose to extract them from the first-order score network via automatic differentiation. We then distill the JVPs into a separate neural network that allows us to efficiently compute the necessary higher-order terms for our novel sampler during synthesis. We only need to train a small additional head on top of the first-order score network. We validate GENIE on multiple image generation benchmarks and demonstrate that GENIE outperforms all previous solvers. Unlike recent methods that fundamentally alter the generation process in DDMs, our GENIE solves the true generative DE and still enables applications such as encoding and guided sampling. Project page and code: https://nv-tlabs.github.io/GENIE.</description><subject>Computer Science - Learning</subject><subject>Statistics - Machine Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotzsEOwUAUheHZWAgewEpfoDUz7a07dlKlEmHBvhnuDJPQyjSEt0dZneRfnHyMDQWPEgTgY-2f7hFJ-Qkckgl02XiZb1b5NCjc6Wx8uPVkfDA3Ve0aV52CubP23ri6Cnb15WF802cdqy-NGfy3x_aLfJ8V4Xq7XGWzdajTCYQKUQlKEQzBgRJAqSjVQkkuDhIRtECQwqQxEmolJQiwPBGWyGg6Kh732Oh324rLm3dX7V_lV1628vgN-zE71w</recordid><startdate>20221011</startdate><enddate>20221011</enddate><creator>Dockhorn, Tim</creator><creator>Vahdat, Arash</creator><creator>Kreis, Karsten</creator><scope>AKY</scope><scope>EPD</scope><scope>GOX</scope></search><sort><creationdate>20221011</creationdate><title>GENIE: Higher-Order Denoising Diffusion Solvers</title><author>Dockhorn, Tim ; Vahdat, Arash ; Kreis, Karsten</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a675-98891d685ed5bd45829d6a19201b2885a18521e638d8a922515f041fddeadc903</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Computer Science - Learning</topic><topic>Statistics - Machine Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Dockhorn, Tim</creatorcontrib><creatorcontrib>Vahdat, Arash</creatorcontrib><creatorcontrib>Kreis, Karsten</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv Statistics</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Dockhorn, Tim</au><au>Vahdat, Arash</au><au>Kreis, Karsten</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>GENIE: Higher-Order Denoising Diffusion Solvers</atitle><date>2022-10-11</date><risdate>2022</risdate><abstract>Denoising diffusion models (DDMs) have emerged as a powerful class of generative models. A forward diffusion process slowly perturbs the data, while a deep model learns to gradually denoise. Synthesis amounts to solving a differential equation (DE) defined by the learnt model. Solving the DE requires slow iterative solvers for high-quality generation. In this work, we propose Higher-Order Denoising Diffusion Solvers (GENIE): Based on truncated Taylor methods, we derive a novel higher-order solver that significantly accelerates synthesis. Our solver relies on higher-order gradients of the perturbed data distribution, that is, higher-order score functions. In practice, only Jacobian-vector products (JVPs) are required and we propose to extract them from the first-order score network via automatic differentiation. We then distill the JVPs into a separate neural network that allows us to efficiently compute the necessary higher-order terms for our novel sampler during synthesis. We only need to train a small additional head on top of the first-order score network. We validate GENIE on multiple image generation benchmarks and demonstrate that GENIE outperforms all previous solvers. Unlike recent methods that fundamentally alter the generation process in DDMs, our GENIE solves the true generative DE and still enables applications such as encoding and guided sampling. Project page and code: https://nv-tlabs.github.io/GENIE.</abstract><doi>10.48550/arxiv.2210.05475</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2210.05475
ispartof
issn
language eng
recordid cdi_arxiv_primary_2210_05475
source arXiv.org
subjects Computer Science - Learning
Statistics - Machine Learning
title GENIE: Higher-Order Denoising Diffusion Solvers
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-08T06%3A44%3A31IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=GENIE:%20Higher-Order%20Denoising%20Diffusion%20Solvers&rft.au=Dockhorn,%20Tim&rft.date=2022-10-11&rft_id=info:doi/10.48550/arxiv.2210.05475&rft_dat=%3Carxiv_GOX%3E2210_05475%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true