Temporal Dynamic Quantization for Diffusion Models
The diffusion model has gained popularity in vision applications due to its remarkable generative performance and versatility. However, high storage and computation demands, resulting from the model size and iterative generation, hinder its use on mobile devices. Existing quantization techniques str...
Gespeichert in:
Hauptverfasser: | , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | So, Junhyuk Lee, Jungwon Ahn, Daehyun Kim, Hyungjun Park, Eunhyeok |
description | The diffusion model has gained popularity in vision applications due to its
remarkable generative performance and versatility. However, high storage and
computation demands, resulting from the model size and iterative generation,
hinder its use on mobile devices. Existing quantization techniques struggle to
maintain performance even in 8-bit precision due to the diffusion model's
unique property of temporal variation in activation. We introduce a novel
quantization method that dynamically adjusts the quantization interval based on
time step information, significantly improving output quality. Unlike
conventional dynamic quantization techniques, our approach has no computational
overhead during inference and is compatible with both post-training
quantization (PTQ) and quantization-aware training (QAT). Our extensive
experiments demonstrate substantial improvements in output quality with the
quantized diffusion model across various datasets. |
doi_str_mv | 10.48550/arxiv.2306.02316 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2306_02316</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2306_02316</sourcerecordid><originalsourceid>FETCH-LOGICAL-a676-bf8ff0358198d5a01e8b503e4cdee4e13a0142250da3f9288df2582f56200ede3</originalsourceid><addsrcrecordid>eNotzs1uwjAQBGBfeqhoH6Cn5gWSrtexMUcELUUCoUq5RwvelSzlBzlQFZ6en3IazRxGn1JvGorSWwsflP7ib4EGXAFotHtWWHG77xM12fzUURt32c-RukM80yH2XSZ9yuZR5Djc2roP3Awv6kmoGfj1kSNVfX1Ws-98tVksZ9NVTm7s8q14ETDW64kPlkCz31owXO4Cc8naXKcS0UIgIxP0Pghaj2IdAnBgM1Lv_7d3dL1PsaV0qm_4-o43F-DwPqQ</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Temporal Dynamic Quantization for Diffusion Models</title><source>arXiv.org</source><creator>So, Junhyuk ; Lee, Jungwon ; Ahn, Daehyun ; Kim, Hyungjun ; Park, Eunhyeok</creator><creatorcontrib>So, Junhyuk ; Lee, Jungwon ; Ahn, Daehyun ; Kim, Hyungjun ; Park, Eunhyeok</creatorcontrib><description>The diffusion model has gained popularity in vision applications due to its
remarkable generative performance and versatility. However, high storage and
computation demands, resulting from the model size and iterative generation,
hinder its use on mobile devices. Existing quantization techniques struggle to
maintain performance even in 8-bit precision due to the diffusion model's
unique property of temporal variation in activation. We introduce a novel
quantization method that dynamically adjusts the quantization interval based on
time step information, significantly improving output quality. Unlike
conventional dynamic quantization techniques, our approach has no computational
overhead during inference and is compatible with both post-training
quantization (PTQ) and quantization-aware training (QAT). Our extensive
experiments demonstrate substantial improvements in output quality with the
quantized diffusion model across various datasets.</description><identifier>DOI: 10.48550/arxiv.2306.02316</identifier><language>eng</language><subject>Computer Science - Computer Vision and Pattern Recognition</subject><creationdate>2023-06</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2306.02316$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2306.02316$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>So, Junhyuk</creatorcontrib><creatorcontrib>Lee, Jungwon</creatorcontrib><creatorcontrib>Ahn, Daehyun</creatorcontrib><creatorcontrib>Kim, Hyungjun</creatorcontrib><creatorcontrib>Park, Eunhyeok</creatorcontrib><title>Temporal Dynamic Quantization for Diffusion Models</title><description>The diffusion model has gained popularity in vision applications due to its
remarkable generative performance and versatility. However, high storage and
computation demands, resulting from the model size and iterative generation,
hinder its use on mobile devices. Existing quantization techniques struggle to
maintain performance even in 8-bit precision due to the diffusion model's
unique property of temporal variation in activation. We introduce a novel
quantization method that dynamically adjusts the quantization interval based on
time step information, significantly improving output quality. Unlike
conventional dynamic quantization techniques, our approach has no computational
overhead during inference and is compatible with both post-training
quantization (PTQ) and quantization-aware training (QAT). Our extensive
experiments demonstrate substantial improvements in output quality with the
quantized diffusion model across various datasets.</description><subject>Computer Science - Computer Vision and Pattern Recognition</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotzs1uwjAQBGBfeqhoH6Cn5gWSrtexMUcELUUCoUq5RwvelSzlBzlQFZ6en3IazRxGn1JvGorSWwsflP7ib4EGXAFotHtWWHG77xM12fzUURt32c-RukM80yH2XSZ9yuZR5Djc2roP3Awv6kmoGfj1kSNVfX1Ws-98tVksZ9NVTm7s8q14ETDW64kPlkCz31owXO4Cc8naXKcS0UIgIxP0Pghaj2IdAnBgM1Lv_7d3dL1PsaV0qm_4-o43F-DwPqQ</recordid><startdate>20230604</startdate><enddate>20230604</enddate><creator>So, Junhyuk</creator><creator>Lee, Jungwon</creator><creator>Ahn, Daehyun</creator><creator>Kim, Hyungjun</creator><creator>Park, Eunhyeok</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20230604</creationdate><title>Temporal Dynamic Quantization for Diffusion Models</title><author>So, Junhyuk ; Lee, Jungwon ; Ahn, Daehyun ; Kim, Hyungjun ; Park, Eunhyeok</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a676-bf8ff0358198d5a01e8b503e4cdee4e13a0142250da3f9288df2582f56200ede3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science - Computer Vision and Pattern Recognition</topic><toplevel>online_resources</toplevel><creatorcontrib>So, Junhyuk</creatorcontrib><creatorcontrib>Lee, Jungwon</creatorcontrib><creatorcontrib>Ahn, Daehyun</creatorcontrib><creatorcontrib>Kim, Hyungjun</creatorcontrib><creatorcontrib>Park, Eunhyeok</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>So, Junhyuk</au><au>Lee, Jungwon</au><au>Ahn, Daehyun</au><au>Kim, Hyungjun</au><au>Park, Eunhyeok</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Temporal Dynamic Quantization for Diffusion Models</atitle><date>2023-06-04</date><risdate>2023</risdate><abstract>The diffusion model has gained popularity in vision applications due to its
remarkable generative performance and versatility. However, high storage and
computation demands, resulting from the model size and iterative generation,
hinder its use on mobile devices. Existing quantization techniques struggle to
maintain performance even in 8-bit precision due to the diffusion model's
unique property of temporal variation in activation. We introduce a novel
quantization method that dynamically adjusts the quantization interval based on
time step information, significantly improving output quality. Unlike
conventional dynamic quantization techniques, our approach has no computational
overhead during inference and is compatible with both post-training
quantization (PTQ) and quantization-aware training (QAT). Our extensive
experiments demonstrate substantial improvements in output quality with the
quantized diffusion model across various datasets.</abstract><doi>10.48550/arxiv.2306.02316</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2306.02316 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2306_02316 |
source | arXiv.org |
subjects | Computer Science - Computer Vision and Pattern Recognition |
title | Temporal Dynamic Quantization for Diffusion Models |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-28T13%3A10%3A28IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Temporal%20Dynamic%20Quantization%20for%20Diffusion%20Models&rft.au=So,%20Junhyuk&rft.date=2023-06-04&rft_id=info:doi/10.48550/arxiv.2306.02316&rft_dat=%3Carxiv_GOX%3E2306_02316%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |