DiffFace: Diffusion-based Face Swapping with Facial Guidance
In this paper, we propose a diffusion-based face swapping framework for the first time, called DiffFace, composed of training ID conditional DDPM, sampling with facial guidance, and a target-preserving blending. In specific, in the training process, the ID conditional DDPM is trained to generate fac...
Gespeichert in:
Veröffentlicht in: | arXiv.org 2022-12 |
---|---|
Hauptverfasser: | , , , , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | arXiv.org |
container_volume | |
creator | Kim, Kihong Kim, Yunho Cho, Seokju Seo, Junyoung Nam, Jisu Lee, Kychul Kim, Seungryong Lee, KwangHee |
description | In this paper, we propose a diffusion-based face swapping framework for the first time, called DiffFace, composed of training ID conditional DDPM, sampling with facial guidance, and a target-preserving blending. In specific, in the training process, the ID conditional DDPM is trained to generate face images with the desired identity. In the sampling process, we use the off-the-shelf facial expert models to make the model transfer source identity while preserving target attributes faithfully. During this process, to preserve the background of the target image and obtain the desired face swapping result, we additionally propose a target-preserving blending strategy. It helps our model to keep the attributes of the target face from noise while transferring the source facial identity. In addition, without any re-training, our model can flexibly apply additional facial guidance and adaptively control the ID-attributes trade-off to achieve the desired results. To the best of our knowledge, this is the first approach that applies the diffusion model in face swapping task. Compared with previous GAN-based approaches, by taking advantage of the diffusion model for the face swapping task, DiffFace achieves better benefits such as training stability, high fidelity, diversity of the samples, and controllability. Extensive experiments show that our DiffFace is comparable or superior to the state-of-the-art methods on several standard face swapping benchmarks. |
format | Article |
fullrecord | <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2759127977</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2759127977</sourcerecordid><originalsourceid>FETCH-proquest_journals_27591279773</originalsourceid><addsrcrecordid>eNpjYuA0MjY21LUwMTLiYOAtLs4yMDAwMjM3MjU15mSwcclMS3NLTE61UgCxSosz8_N0kxKLU1MUQKIKweWJBQWZeekK5ZklGSChzMQcBffSzJTEvORUHgbWtMSc4lReKM3NoOzmGuLsoVtQlF9YmlpcEp-VX1qUB5SKNzI3tTQ0Mrc0NzcmThUAxTI2ug</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2759127977</pqid></control><display><type>article</type><title>DiffFace: Diffusion-based Face Swapping with Facial Guidance</title><source>Freely Accessible Journals</source><creator>Kim, Kihong ; Kim, Yunho ; Cho, Seokju ; Seo, Junyoung ; Nam, Jisu ; Lee, Kychul ; Kim, Seungryong ; Lee, KwangHee</creator><creatorcontrib>Kim, Kihong ; Kim, Yunho ; Cho, Seokju ; Seo, Junyoung ; Nam, Jisu ; Lee, Kychul ; Kim, Seungryong ; Lee, KwangHee</creatorcontrib><description>In this paper, we propose a diffusion-based face swapping framework for the first time, called DiffFace, composed of training ID conditional DDPM, sampling with facial guidance, and a target-preserving blending. In specific, in the training process, the ID conditional DDPM is trained to generate face images with the desired identity. In the sampling process, we use the off-the-shelf facial expert models to make the model transfer source identity while preserving target attributes faithfully. During this process, to preserve the background of the target image and obtain the desired face swapping result, we additionally propose a target-preserving blending strategy. It helps our model to keep the attributes of the target face from noise while transferring the source facial identity. In addition, without any re-training, our model can flexibly apply additional facial guidance and adaptively control the ID-attributes trade-off to achieve the desired results. To the best of our knowledge, this is the first approach that applies the diffusion model in face swapping task. Compared with previous GAN-based approaches, by taking advantage of the diffusion model for the face swapping task, DiffFace achieves better benefits such as training stability, high fidelity, diversity of the samples, and controllability. Extensive experiments show that our DiffFace is comparable or superior to the state-of-the-art methods on several standard face swapping benchmarks.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Blending ; Diffusion ; Sampling ; Training</subject><ispartof>arXiv.org, 2022-12</ispartof><rights>2022. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>781,785</link.rule.ids></links><search><creatorcontrib>Kim, Kihong</creatorcontrib><creatorcontrib>Kim, Yunho</creatorcontrib><creatorcontrib>Cho, Seokju</creatorcontrib><creatorcontrib>Seo, Junyoung</creatorcontrib><creatorcontrib>Nam, Jisu</creatorcontrib><creatorcontrib>Lee, Kychul</creatorcontrib><creatorcontrib>Kim, Seungryong</creatorcontrib><creatorcontrib>Lee, KwangHee</creatorcontrib><title>DiffFace: Diffusion-based Face Swapping with Facial Guidance</title><title>arXiv.org</title><description>In this paper, we propose a diffusion-based face swapping framework for the first time, called DiffFace, composed of training ID conditional DDPM, sampling with facial guidance, and a target-preserving blending. In specific, in the training process, the ID conditional DDPM is trained to generate face images with the desired identity. In the sampling process, we use the off-the-shelf facial expert models to make the model transfer source identity while preserving target attributes faithfully. During this process, to preserve the background of the target image and obtain the desired face swapping result, we additionally propose a target-preserving blending strategy. It helps our model to keep the attributes of the target face from noise while transferring the source facial identity. In addition, without any re-training, our model can flexibly apply additional facial guidance and adaptively control the ID-attributes trade-off to achieve the desired results. To the best of our knowledge, this is the first approach that applies the diffusion model in face swapping task. Compared with previous GAN-based approaches, by taking advantage of the diffusion model for the face swapping task, DiffFace achieves better benefits such as training stability, high fidelity, diversity of the samples, and controllability. Extensive experiments show that our DiffFace is comparable or superior to the state-of-the-art methods on several standard face swapping benchmarks.</description><subject>Blending</subject><subject>Diffusion</subject><subject>Sampling</subject><subject>Training</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><recordid>eNpjYuA0MjY21LUwMTLiYOAtLs4yMDAwMjM3MjU15mSwcclMS3NLTE61UgCxSosz8_N0kxKLU1MUQKIKweWJBQWZeekK5ZklGSChzMQcBffSzJTEvORUHgbWtMSc4lReKM3NoOzmGuLsoVtQlF9YmlpcEp-VX1qUB5SKNzI3tTQ0Mrc0NzcmThUAxTI2ug</recordid><startdate>20221227</startdate><enddate>20221227</enddate><creator>Kim, Kihong</creator><creator>Kim, Yunho</creator><creator>Cho, Seokju</creator><creator>Seo, Junyoung</creator><creator>Nam, Jisu</creator><creator>Lee, Kychul</creator><creator>Kim, Seungryong</creator><creator>Lee, KwangHee</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20221227</creationdate><title>DiffFace: Diffusion-based Face Swapping with Facial Guidance</title><author>Kim, Kihong ; Kim, Yunho ; Cho, Seokju ; Seo, Junyoung ; Nam, Jisu ; Lee, Kychul ; Kim, Seungryong ; Lee, KwangHee</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_27591279773</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Blending</topic><topic>Diffusion</topic><topic>Sampling</topic><topic>Training</topic><toplevel>online_resources</toplevel><creatorcontrib>Kim, Kihong</creatorcontrib><creatorcontrib>Kim, Yunho</creatorcontrib><creatorcontrib>Cho, Seokju</creatorcontrib><creatorcontrib>Seo, Junyoung</creatorcontrib><creatorcontrib>Nam, Jisu</creatorcontrib><creatorcontrib>Lee, Kychul</creatorcontrib><creatorcontrib>Kim, Seungryong</creatorcontrib><creatorcontrib>Lee, KwangHee</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Access via ProQuest (Open Access)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Kim, Kihong</au><au>Kim, Yunho</au><au>Cho, Seokju</au><au>Seo, Junyoung</au><au>Nam, Jisu</au><au>Lee, Kychul</au><au>Kim, Seungryong</au><au>Lee, KwangHee</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>DiffFace: Diffusion-based Face Swapping with Facial Guidance</atitle><jtitle>arXiv.org</jtitle><date>2022-12-27</date><risdate>2022</risdate><eissn>2331-8422</eissn><abstract>In this paper, we propose a diffusion-based face swapping framework for the first time, called DiffFace, composed of training ID conditional DDPM, sampling with facial guidance, and a target-preserving blending. In specific, in the training process, the ID conditional DDPM is trained to generate face images with the desired identity. In the sampling process, we use the off-the-shelf facial expert models to make the model transfer source identity while preserving target attributes faithfully. During this process, to preserve the background of the target image and obtain the desired face swapping result, we additionally propose a target-preserving blending strategy. It helps our model to keep the attributes of the target face from noise while transferring the source facial identity. In addition, without any re-training, our model can flexibly apply additional facial guidance and adaptively control the ID-attributes trade-off to achieve the desired results. To the best of our knowledge, this is the first approach that applies the diffusion model in face swapping task. Compared with previous GAN-based approaches, by taking advantage of the diffusion model for the face swapping task, DiffFace achieves better benefits such as training stability, high fidelity, diversity of the samples, and controllability. Extensive experiments show that our DiffFace is comparable or superior to the state-of-the-art methods on several standard face swapping benchmarks.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | EISSN: 2331-8422 |
ispartof | arXiv.org, 2022-12 |
issn | 2331-8422 |
language | eng |
recordid | cdi_proquest_journals_2759127977 |
source | Freely Accessible Journals |
subjects | Blending Diffusion Sampling Training |
title | DiffFace: Diffusion-based Face Swapping with Facial Guidance |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-16T04%3A32%3A59IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=DiffFace:%20Diffusion-based%20Face%20Swapping%20with%20Facial%20Guidance&rft.jtitle=arXiv.org&rft.au=Kim,%20Kihong&rft.date=2022-12-27&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2759127977%3C/proquest%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2759127977&rft_id=info:pmid/&rfr_iscdi=true |