Temporal Dynamic Quantization for Diffusion Models

The diffusion model has gained popularity in vision applications due to its remarkable generative performance and versatility. However, high storage and computation demands, resulting from the model size and iterative generation, hinder its use on mobile devices. Existing quantization techniques str...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: So, Junhyuk, Lee, Jungwon, Ahn, Daehyun, Kim, Hyungjun, Park, Eunhyeok
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator So, Junhyuk
Lee, Jungwon
Ahn, Daehyun
Kim, Hyungjun
Park, Eunhyeok
description The diffusion model has gained popularity in vision applications due to its remarkable generative performance and versatility. However, high storage and computation demands, resulting from the model size and iterative generation, hinder its use on mobile devices. Existing quantization techniques struggle to maintain performance even in 8-bit precision due to the diffusion model's unique property of temporal variation in activation. We introduce a novel quantization method that dynamically adjusts the quantization interval based on time step information, significantly improving output quality. Unlike conventional dynamic quantization techniques, our approach has no computational overhead during inference and is compatible with both post-training quantization (PTQ) and quantization-aware training (QAT). Our extensive experiments demonstrate substantial improvements in output quality with the quantized diffusion model across various datasets.
doi_str_mv 10.48550/arxiv.2306.02316
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2306_02316</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2306_02316</sourcerecordid><originalsourceid>FETCH-LOGICAL-a676-bf8ff0358198d5a01e8b503e4cdee4e13a0142250da3f9288df2582f56200ede3</originalsourceid><addsrcrecordid>eNotzs1uwjAQBGBfeqhoH6Cn5gWSrtexMUcELUUCoUq5RwvelSzlBzlQFZ6en3IazRxGn1JvGorSWwsflP7ib4EGXAFotHtWWHG77xM12fzUURt32c-RukM80yH2XSZ9yuZR5Djc2roP3Awv6kmoGfj1kSNVfX1Ws-98tVksZ9NVTm7s8q14ETDW64kPlkCz31owXO4Cc8naXKcS0UIgIxP0Pghaj2IdAnBgM1Lv_7d3dL1PsaV0qm_4-o43F-DwPqQ</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Temporal Dynamic Quantization for Diffusion Models</title><source>arXiv.org</source><creator>So, Junhyuk ; Lee, Jungwon ; Ahn, Daehyun ; Kim, Hyungjun ; Park, Eunhyeok</creator><creatorcontrib>So, Junhyuk ; Lee, Jungwon ; Ahn, Daehyun ; Kim, Hyungjun ; Park, Eunhyeok</creatorcontrib><description>The diffusion model has gained popularity in vision applications due to its remarkable generative performance and versatility. However, high storage and computation demands, resulting from the model size and iterative generation, hinder its use on mobile devices. Existing quantization techniques struggle to maintain performance even in 8-bit precision due to the diffusion model's unique property of temporal variation in activation. We introduce a novel quantization method that dynamically adjusts the quantization interval based on time step information, significantly improving output quality. Unlike conventional dynamic quantization techniques, our approach has no computational overhead during inference and is compatible with both post-training quantization (PTQ) and quantization-aware training (QAT). Our extensive experiments demonstrate substantial improvements in output quality with the quantized diffusion model across various datasets.</description><identifier>DOI: 10.48550/arxiv.2306.02316</identifier><language>eng</language><subject>Computer Science - Computer Vision and Pattern Recognition</subject><creationdate>2023-06</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2306.02316$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2306.02316$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>So, Junhyuk</creatorcontrib><creatorcontrib>Lee, Jungwon</creatorcontrib><creatorcontrib>Ahn, Daehyun</creatorcontrib><creatorcontrib>Kim, Hyungjun</creatorcontrib><creatorcontrib>Park, Eunhyeok</creatorcontrib><title>Temporal Dynamic Quantization for Diffusion Models</title><description>The diffusion model has gained popularity in vision applications due to its remarkable generative performance and versatility. However, high storage and computation demands, resulting from the model size and iterative generation, hinder its use on mobile devices. Existing quantization techniques struggle to maintain performance even in 8-bit precision due to the diffusion model's unique property of temporal variation in activation. We introduce a novel quantization method that dynamically adjusts the quantization interval based on time step information, significantly improving output quality. Unlike conventional dynamic quantization techniques, our approach has no computational overhead during inference and is compatible with both post-training quantization (PTQ) and quantization-aware training (QAT). Our extensive experiments demonstrate substantial improvements in output quality with the quantized diffusion model across various datasets.</description><subject>Computer Science - Computer Vision and Pattern Recognition</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotzs1uwjAQBGBfeqhoH6Cn5gWSrtexMUcELUUCoUq5RwvelSzlBzlQFZ6en3IazRxGn1JvGorSWwsflP7ib4EGXAFotHtWWHG77xM12fzUURt32c-RukM80yH2XSZ9yuZR5Djc2roP3Awv6kmoGfj1kSNVfX1Ws-98tVksZ9NVTm7s8q14ETDW64kPlkCz31owXO4Cc8naXKcS0UIgIxP0Pghaj2IdAnBgM1Lv_7d3dL1PsaV0qm_4-o43F-DwPqQ</recordid><startdate>20230604</startdate><enddate>20230604</enddate><creator>So, Junhyuk</creator><creator>Lee, Jungwon</creator><creator>Ahn, Daehyun</creator><creator>Kim, Hyungjun</creator><creator>Park, Eunhyeok</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20230604</creationdate><title>Temporal Dynamic Quantization for Diffusion Models</title><author>So, Junhyuk ; Lee, Jungwon ; Ahn, Daehyun ; Kim, Hyungjun ; Park, Eunhyeok</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a676-bf8ff0358198d5a01e8b503e4cdee4e13a0142250da3f9288df2582f56200ede3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science - Computer Vision and Pattern Recognition</topic><toplevel>online_resources</toplevel><creatorcontrib>So, Junhyuk</creatorcontrib><creatorcontrib>Lee, Jungwon</creatorcontrib><creatorcontrib>Ahn, Daehyun</creatorcontrib><creatorcontrib>Kim, Hyungjun</creatorcontrib><creatorcontrib>Park, Eunhyeok</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>So, Junhyuk</au><au>Lee, Jungwon</au><au>Ahn, Daehyun</au><au>Kim, Hyungjun</au><au>Park, Eunhyeok</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Temporal Dynamic Quantization for Diffusion Models</atitle><date>2023-06-04</date><risdate>2023</risdate><abstract>The diffusion model has gained popularity in vision applications due to its remarkable generative performance and versatility. However, high storage and computation demands, resulting from the model size and iterative generation, hinder its use on mobile devices. Existing quantization techniques struggle to maintain performance even in 8-bit precision due to the diffusion model's unique property of temporal variation in activation. We introduce a novel quantization method that dynamically adjusts the quantization interval based on time step information, significantly improving output quality. Unlike conventional dynamic quantization techniques, our approach has no computational overhead during inference and is compatible with both post-training quantization (PTQ) and quantization-aware training (QAT). Our extensive experiments demonstrate substantial improvements in output quality with the quantized diffusion model across various datasets.</abstract><doi>10.48550/arxiv.2306.02316</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2306.02316
ispartof
issn
language eng
recordid cdi_arxiv_primary_2306_02316
source arXiv.org
subjects Computer Science - Computer Vision and Pattern Recognition
title Temporal Dynamic Quantization for Diffusion Models
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-28T13%3A10%3A28IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Temporal%20Dynamic%20Quantization%20for%20Diffusion%20Models&rft.au=So,%20Junhyuk&rft.date=2023-06-04&rft_id=info:doi/10.48550/arxiv.2306.02316&rft_dat=%3Carxiv_GOX%3E2306_02316%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true