DiffFace: Diffusion-based Face Swapping with Facial Guidance

In this paper, we propose a diffusion-based face swapping framework for the first time, called DiffFace, composed of training ID conditional DDPM, sampling with facial guidance, and a target-preserving blending. In specific, in the training process, the ID conditional DDPM is trained to generate fac...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:arXiv.org 2022-12
Hauptverfasser: Kim, Kihong, Kim, Yunho, Cho, Seokju, Seo, Junyoung, Nam, Jisu, Lee, Kychul, Kim, Seungryong, Lee, KwangHee
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Kim, Kihong
Kim, Yunho
Cho, Seokju
Seo, Junyoung
Nam, Jisu
Lee, Kychul
Kim, Seungryong
Lee, KwangHee
description In this paper, we propose a diffusion-based face swapping framework for the first time, called DiffFace, composed of training ID conditional DDPM, sampling with facial guidance, and a target-preserving blending. In specific, in the training process, the ID conditional DDPM is trained to generate face images with the desired identity. In the sampling process, we use the off-the-shelf facial expert models to make the model transfer source identity while preserving target attributes faithfully. During this process, to preserve the background of the target image and obtain the desired face swapping result, we additionally propose a target-preserving blending strategy. It helps our model to keep the attributes of the target face from noise while transferring the source facial identity. In addition, without any re-training, our model can flexibly apply additional facial guidance and adaptively control the ID-attributes trade-off to achieve the desired results. To the best of our knowledge, this is the first approach that applies the diffusion model in face swapping task. Compared with previous GAN-based approaches, by taking advantage of the diffusion model for the face swapping task, DiffFace achieves better benefits such as training stability, high fidelity, diversity of the samples, and controllability. Extensive experiments show that our DiffFace is comparable or superior to the state-of-the-art methods on several standard face swapping benchmarks.
format Article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2759127977</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2759127977</sourcerecordid><originalsourceid>FETCH-proquest_journals_27591279773</originalsourceid><addsrcrecordid>eNpjYuA0MjY21LUwMTLiYOAtLs4yMDAwMjM3MjU15mSwcclMS3NLTE61UgCxSosz8_N0kxKLU1MUQKIKweWJBQWZeekK5ZklGSChzMQcBffSzJTEvORUHgbWtMSc4lReKM3NoOzmGuLsoVtQlF9YmlpcEp-VX1qUB5SKNzI3tTQ0Mrc0NzcmThUAxTI2ug</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2759127977</pqid></control><display><type>article</type><title>DiffFace: Diffusion-based Face Swapping with Facial Guidance</title><source>Freely Accessible Journals</source><creator>Kim, Kihong ; Kim, Yunho ; Cho, Seokju ; Seo, Junyoung ; Nam, Jisu ; Lee, Kychul ; Kim, Seungryong ; Lee, KwangHee</creator><creatorcontrib>Kim, Kihong ; Kim, Yunho ; Cho, Seokju ; Seo, Junyoung ; Nam, Jisu ; Lee, Kychul ; Kim, Seungryong ; Lee, KwangHee</creatorcontrib><description>In this paper, we propose a diffusion-based face swapping framework for the first time, called DiffFace, composed of training ID conditional DDPM, sampling with facial guidance, and a target-preserving blending. In specific, in the training process, the ID conditional DDPM is trained to generate face images with the desired identity. In the sampling process, we use the off-the-shelf facial expert models to make the model transfer source identity while preserving target attributes faithfully. During this process, to preserve the background of the target image and obtain the desired face swapping result, we additionally propose a target-preserving blending strategy. It helps our model to keep the attributes of the target face from noise while transferring the source facial identity. In addition, without any re-training, our model can flexibly apply additional facial guidance and adaptively control the ID-attributes trade-off to achieve the desired results. To the best of our knowledge, this is the first approach that applies the diffusion model in face swapping task. Compared with previous GAN-based approaches, by taking advantage of the diffusion model for the face swapping task, DiffFace achieves better benefits such as training stability, high fidelity, diversity of the samples, and controllability. Extensive experiments show that our DiffFace is comparable or superior to the state-of-the-art methods on several standard face swapping benchmarks.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Blending ; Diffusion ; Sampling ; Training</subject><ispartof>arXiv.org, 2022-12</ispartof><rights>2022. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>781,785</link.rule.ids></links><search><creatorcontrib>Kim, Kihong</creatorcontrib><creatorcontrib>Kim, Yunho</creatorcontrib><creatorcontrib>Cho, Seokju</creatorcontrib><creatorcontrib>Seo, Junyoung</creatorcontrib><creatorcontrib>Nam, Jisu</creatorcontrib><creatorcontrib>Lee, Kychul</creatorcontrib><creatorcontrib>Kim, Seungryong</creatorcontrib><creatorcontrib>Lee, KwangHee</creatorcontrib><title>DiffFace: Diffusion-based Face Swapping with Facial Guidance</title><title>arXiv.org</title><description>In this paper, we propose a diffusion-based face swapping framework for the first time, called DiffFace, composed of training ID conditional DDPM, sampling with facial guidance, and a target-preserving blending. In specific, in the training process, the ID conditional DDPM is trained to generate face images with the desired identity. In the sampling process, we use the off-the-shelf facial expert models to make the model transfer source identity while preserving target attributes faithfully. During this process, to preserve the background of the target image and obtain the desired face swapping result, we additionally propose a target-preserving blending strategy. It helps our model to keep the attributes of the target face from noise while transferring the source facial identity. In addition, without any re-training, our model can flexibly apply additional facial guidance and adaptively control the ID-attributes trade-off to achieve the desired results. To the best of our knowledge, this is the first approach that applies the diffusion model in face swapping task. Compared with previous GAN-based approaches, by taking advantage of the diffusion model for the face swapping task, DiffFace achieves better benefits such as training stability, high fidelity, diversity of the samples, and controllability. Extensive experiments show that our DiffFace is comparable or superior to the state-of-the-art methods on several standard face swapping benchmarks.</description><subject>Blending</subject><subject>Diffusion</subject><subject>Sampling</subject><subject>Training</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><recordid>eNpjYuA0MjY21LUwMTLiYOAtLs4yMDAwMjM3MjU15mSwcclMS3NLTE61UgCxSosz8_N0kxKLU1MUQKIKweWJBQWZeekK5ZklGSChzMQcBffSzJTEvORUHgbWtMSc4lReKM3NoOzmGuLsoVtQlF9YmlpcEp-VX1qUB5SKNzI3tTQ0Mrc0NzcmThUAxTI2ug</recordid><startdate>20221227</startdate><enddate>20221227</enddate><creator>Kim, Kihong</creator><creator>Kim, Yunho</creator><creator>Cho, Seokju</creator><creator>Seo, Junyoung</creator><creator>Nam, Jisu</creator><creator>Lee, Kychul</creator><creator>Kim, Seungryong</creator><creator>Lee, KwangHee</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20221227</creationdate><title>DiffFace: Diffusion-based Face Swapping with Facial Guidance</title><author>Kim, Kihong ; Kim, Yunho ; Cho, Seokju ; Seo, Junyoung ; Nam, Jisu ; Lee, Kychul ; Kim, Seungryong ; Lee, KwangHee</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_27591279773</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Blending</topic><topic>Diffusion</topic><topic>Sampling</topic><topic>Training</topic><toplevel>online_resources</toplevel><creatorcontrib>Kim, Kihong</creatorcontrib><creatorcontrib>Kim, Yunho</creatorcontrib><creatorcontrib>Cho, Seokju</creatorcontrib><creatorcontrib>Seo, Junyoung</creatorcontrib><creatorcontrib>Nam, Jisu</creatorcontrib><creatorcontrib>Lee, Kychul</creatorcontrib><creatorcontrib>Kim, Seungryong</creatorcontrib><creatorcontrib>Lee, KwangHee</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Access via ProQuest (Open Access)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Kim, Kihong</au><au>Kim, Yunho</au><au>Cho, Seokju</au><au>Seo, Junyoung</au><au>Nam, Jisu</au><au>Lee, Kychul</au><au>Kim, Seungryong</au><au>Lee, KwangHee</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>DiffFace: Diffusion-based Face Swapping with Facial Guidance</atitle><jtitle>arXiv.org</jtitle><date>2022-12-27</date><risdate>2022</risdate><eissn>2331-8422</eissn><abstract>In this paper, we propose a diffusion-based face swapping framework for the first time, called DiffFace, composed of training ID conditional DDPM, sampling with facial guidance, and a target-preserving blending. In specific, in the training process, the ID conditional DDPM is trained to generate face images with the desired identity. In the sampling process, we use the off-the-shelf facial expert models to make the model transfer source identity while preserving target attributes faithfully. During this process, to preserve the background of the target image and obtain the desired face swapping result, we additionally propose a target-preserving blending strategy. It helps our model to keep the attributes of the target face from noise while transferring the source facial identity. In addition, without any re-training, our model can flexibly apply additional facial guidance and adaptively control the ID-attributes trade-off to achieve the desired results. To the best of our knowledge, this is the first approach that applies the diffusion model in face swapping task. Compared with previous GAN-based approaches, by taking advantage of the diffusion model for the face swapping task, DiffFace achieves better benefits such as training stability, high fidelity, diversity of the samples, and controllability. Extensive experiments show that our DiffFace is comparable or superior to the state-of-the-art methods on several standard face swapping benchmarks.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2022-12
issn 2331-8422
language eng
recordid cdi_proquest_journals_2759127977
source Freely Accessible Journals
subjects Blending
Diffusion
Sampling
Training
title DiffFace: Diffusion-based Face Swapping with Facial Guidance
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-16T04%3A32%3A59IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=DiffFace:%20Diffusion-based%20Face%20Swapping%20with%20Facial%20Guidance&rft.jtitle=arXiv.org&rft.au=Kim,%20Kihong&rft.date=2022-12-27&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2759127977%3C/proquest%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2759127977&rft_id=info:pmid/&rfr_iscdi=true