DEEPFAKE Image Synthesis for Data Augmentation

Field of medical imaging is scarce in terms of a dataset that is reliable and extensive enough to train distinct supervised deep learning models. One way to tackle this problem is to use a Generative Adversarial Network to synthesize DEEPFAKE images to augment the data. DEEPFAKE refers to the transf...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE access 2022, Vol.10, p.80847-80857
Hauptverfasser: Waqas, Nawaf, Safie, Sairul Izwan, Kadir, Kushsairy Abdul, Khan, Sheroz, Kaka Khel, Muhammad Haris
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 80857
container_issue
container_start_page 80847
container_title IEEE access
container_volume 10
creator Waqas, Nawaf
Safie, Sairul Izwan
Kadir, Kushsairy Abdul
Khan, Sheroz
Kaka Khel, Muhammad Haris
description Field of medical imaging is scarce in terms of a dataset that is reliable and extensive enough to train distinct supervised deep learning models. One way to tackle this problem is to use a Generative Adversarial Network to synthesize DEEPFAKE images to augment the data. DEEPFAKE refers to the transfer of important features from the source image (or video) to the target image (or video), such that the target modality appears to animate the source almost close to reality. In the past decade, medical image processing has made significant advances using the latest state-of-art-methods of deep learning techniques. Supervised deep learning models produce super-human results with the help of huge amount of dataset in a variety of medical image processing and deep learning applications. DEEPFAKE images can be a useful in various applications like translating to different useful and sometimes malicious modalities, unbalanced datasets or increasing the amount of datasets. In this paper the data scarcity has been addressed by using Progressive Growing Generative Adversarial Networks (PGGAN). However, PGGAN consists of convolution layer that suffers from the training-related issues. PGGAN requires a large number of convolution layers in order to obtain high-resolution image training, which makes training a difficult task. In this work, a subjective self-attention layer has been added before 256 \times 256 convolution layer for efficient feature learning and the use of spectral normalization in the discriminator and pixel normalization in the generator for training stabilization - the two tasks resulting into what is referred to as Enhanced-GAN. The performance of Enhanced-GAN is compared to PGGAN performance using the parameters of AM Score and Mode Score. In addition, the strength of Enhanced-GAN and PGGAN synthesized data is evaluated using the U-net supervised deep learning model for segmentation tasks. Dice Coefficient metrics show that U-net trained on Enhanced-GAN DEEPFAKE data optimized with real data performs better than PGGAN DEEPFAKE data with real data.
doi_str_mv 10.1109/ACCESS.2022.3193668
format Article
fullrecord <record><control><sourceid>proquest_doaj_</sourceid><recordid>TN_cdi_proquest_journals_2700412434</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9839427</ieee_id><doaj_id>oai_doaj_org_article_368243818ee641de85ce183a2bf84e65</doaj_id><sourcerecordid>2700412434</sourcerecordid><originalsourceid>FETCH-LOGICAL-c408t-f725ac32b6e75d605a57f8beb6094a805003507b06ce76da52ecf2a993ef30cc3</originalsourceid><addsrcrecordid>eNpNkE1Lw0AQhoMoWGp_QS8Bz4n7kf06hjTVYkGhel42m0lNabN1Nz3035uaIs5lhpd53xmeKJpjlGKM1FNeFOVmkxJESEqxopzLm2hCMFcJZZTf_pvvo1kIOzSUHCQmJlG6KMv3Zf5axquD2UK8OXf9F4Q2xI3z8cL0Js5P2wN0velb1z1Ed43ZB5hd-zT6XJYfxUuyfnteFfk6sRmSfdIIwoylpOIgWM0RM0w0soKKI5UZiRhClCFRIW5B8NowArYhRikKDUXW0mm0GnNrZ3b66NuD8WftTKt_Bee32vi-tXvQlEuSUYklAM9wDZJZwJIaUjUyA86GrMcx6-jd9wlCr3fu5LvhfU0EQhke7NmwRcct610IHpq_qxjpC2c9ctYXzvrKeXDNR1cLAH8OJanKiKA_3ot15A</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2700412434</pqid></control><display><type>article</type><title>DEEPFAKE Image Synthesis for Data Augmentation</title><source>IEEE Open Access Journals</source><source>DOAJ Directory of Open Access Journals</source><source>Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals</source><creator>Waqas, Nawaf ; Safie, Sairul Izwan ; Kadir, Kushsairy Abdul ; Khan, Sheroz ; Kaka Khel, Muhammad Haris</creator><creatorcontrib>Waqas, Nawaf ; Safie, Sairul Izwan ; Kadir, Kushsairy Abdul ; Khan, Sheroz ; Kaka Khel, Muhammad Haris</creatorcontrib><description>Field of medical imaging is scarce in terms of a dataset that is reliable and extensive enough to train distinct supervised deep learning models. One way to tackle this problem is to use a Generative Adversarial Network to synthesize DEEPFAKE images to augment the data. DEEPFAKE refers to the transfer of important features from the source image (or video) to the target image (or video), such that the target modality appears to animate the source almost close to reality. In the past decade, medical image processing has made significant advances using the latest state-of-art-methods of deep learning techniques. Supervised deep learning models produce super-human results with the help of huge amount of dataset in a variety of medical image processing and deep learning applications. DEEPFAKE images can be a useful in various applications like translating to different useful and sometimes malicious modalities, unbalanced datasets or increasing the amount of datasets. In this paper the data scarcity has been addressed by using Progressive Growing Generative Adversarial Networks (PGGAN). However, PGGAN consists of convolution layer that suffers from the training-related issues. PGGAN requires a large number of convolution layers in order to obtain high-resolution image training, which makes training a difficult task. In this work, a subjective self-attention layer has been added before &lt;inline-formula&gt; &lt;tex-math notation="LaTeX"&gt;256 \times 256 &lt;/tex-math&gt;&lt;/inline-formula&gt; convolution layer for efficient feature learning and the use of spectral normalization in the discriminator and pixel normalization in the generator for training stabilization - the two tasks resulting into what is referred to as Enhanced-GAN. The performance of Enhanced-GAN is compared to PGGAN performance using the parameters of AM Score and Mode Score. In addition, the strength of Enhanced-GAN and PGGAN synthesized data is evaluated using the U-net supervised deep learning model for segmentation tasks. Dice Coefficient metrics show that U-net trained on Enhanced-GAN DEEPFAKE data optimized with real data performs better than PGGAN DEEPFAKE data with real data.</description><identifier>ISSN: 2169-3536</identifier><identifier>EISSN: 2169-3536</identifier><identifier>DOI: 10.1109/ACCESS.2022.3193668</identifier><identifier>CODEN: IAECCG</identifier><language>eng</language><publisher>Piscataway: IEEE</publisher><subject>Biomedical imaging ; Convolution ; Convolutional neural networks ; Data augmentation ; Data models ; Datasets ; Deception ; Deep learning ; DEEPFAKE ; Deepfakes ; Generative adversarial networks ; Image processing ; Image resolution ; Image segmentation ; Medical imaging ; PGGAN ; self-attention layer ; spectral normalization ; Synthesis ; Training ; Training data ; unbalanced dataset</subject><ispartof>IEEE access, 2022, Vol.10, p.80847-80857</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2022</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c408t-f725ac32b6e75d605a57f8beb6094a805003507b06ce76da52ecf2a993ef30cc3</citedby><cites>FETCH-LOGICAL-c408t-f725ac32b6e75d605a57f8beb6094a805003507b06ce76da52ecf2a993ef30cc3</cites><orcidid>0000-0001-5587-4766 ; 0000-0002-5749-8538 ; 0000-0003-1079-5014 ; 0000-0002-5121-3044</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9839427$$EHTML$$P50$$Gieee$$Hfree_for_read</linktohtml><link.rule.ids>314,776,780,860,2096,4010,27610,27900,27901,27902,54908</link.rule.ids></links><search><creatorcontrib>Waqas, Nawaf</creatorcontrib><creatorcontrib>Safie, Sairul Izwan</creatorcontrib><creatorcontrib>Kadir, Kushsairy Abdul</creatorcontrib><creatorcontrib>Khan, Sheroz</creatorcontrib><creatorcontrib>Kaka Khel, Muhammad Haris</creatorcontrib><title>DEEPFAKE Image Synthesis for Data Augmentation</title><title>IEEE access</title><addtitle>Access</addtitle><description>Field of medical imaging is scarce in terms of a dataset that is reliable and extensive enough to train distinct supervised deep learning models. One way to tackle this problem is to use a Generative Adversarial Network to synthesize DEEPFAKE images to augment the data. DEEPFAKE refers to the transfer of important features from the source image (or video) to the target image (or video), such that the target modality appears to animate the source almost close to reality. In the past decade, medical image processing has made significant advances using the latest state-of-art-methods of deep learning techniques. Supervised deep learning models produce super-human results with the help of huge amount of dataset in a variety of medical image processing and deep learning applications. DEEPFAKE images can be a useful in various applications like translating to different useful and sometimes malicious modalities, unbalanced datasets or increasing the amount of datasets. In this paper the data scarcity has been addressed by using Progressive Growing Generative Adversarial Networks (PGGAN). However, PGGAN consists of convolution layer that suffers from the training-related issues. PGGAN requires a large number of convolution layers in order to obtain high-resolution image training, which makes training a difficult task. In this work, a subjective self-attention layer has been added before &lt;inline-formula&gt; &lt;tex-math notation="LaTeX"&gt;256 \times 256 &lt;/tex-math&gt;&lt;/inline-formula&gt; convolution layer for efficient feature learning and the use of spectral normalization in the discriminator and pixel normalization in the generator for training stabilization - the two tasks resulting into what is referred to as Enhanced-GAN. The performance of Enhanced-GAN is compared to PGGAN performance using the parameters of AM Score and Mode Score. In addition, the strength of Enhanced-GAN and PGGAN synthesized data is evaluated using the U-net supervised deep learning model for segmentation tasks. Dice Coefficient metrics show that U-net trained on Enhanced-GAN DEEPFAKE data optimized with real data performs better than PGGAN DEEPFAKE data with real data.</description><subject>Biomedical imaging</subject><subject>Convolution</subject><subject>Convolutional neural networks</subject><subject>Data augmentation</subject><subject>Data models</subject><subject>Datasets</subject><subject>Deception</subject><subject>Deep learning</subject><subject>DEEPFAKE</subject><subject>Deepfakes</subject><subject>Generative adversarial networks</subject><subject>Image processing</subject><subject>Image resolution</subject><subject>Image segmentation</subject><subject>Medical imaging</subject><subject>PGGAN</subject><subject>self-attention layer</subject><subject>spectral normalization</subject><subject>Synthesis</subject><subject>Training</subject><subject>Training data</subject><subject>unbalanced dataset</subject><issn>2169-3536</issn><issn>2169-3536</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>ESBDL</sourceid><sourceid>RIE</sourceid><sourceid>DOA</sourceid><recordid>eNpNkE1Lw0AQhoMoWGp_QS8Bz4n7kf06hjTVYkGhel42m0lNabN1Nz3035uaIs5lhpd53xmeKJpjlGKM1FNeFOVmkxJESEqxopzLm2hCMFcJZZTf_pvvo1kIOzSUHCQmJlG6KMv3Zf5axquD2UK8OXf9F4Q2xI3z8cL0Js5P2wN0velb1z1Ed43ZB5hd-zT6XJYfxUuyfnteFfk6sRmSfdIIwoylpOIgWM0RM0w0soKKI5UZiRhClCFRIW5B8NowArYhRikKDUXW0mm0GnNrZ3b66NuD8WftTKt_Bee32vi-tXvQlEuSUYklAM9wDZJZwJIaUjUyA86GrMcx6-jd9wlCr3fu5LvhfU0EQhke7NmwRcct610IHpq_qxjpC2c9ctYXzvrKeXDNR1cLAH8OJanKiKA_3ot15A</recordid><startdate>2022</startdate><enddate>2022</enddate><creator>Waqas, Nawaf</creator><creator>Safie, Sairul Izwan</creator><creator>Kadir, Kushsairy Abdul</creator><creator>Khan, Sheroz</creator><creator>Kaka Khel, Muhammad Haris</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>ESBDL</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>7SR</scope><scope>8BQ</scope><scope>8FD</scope><scope>JG9</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>DOA</scope><orcidid>https://orcid.org/0000-0001-5587-4766</orcidid><orcidid>https://orcid.org/0000-0002-5749-8538</orcidid><orcidid>https://orcid.org/0000-0003-1079-5014</orcidid><orcidid>https://orcid.org/0000-0002-5121-3044</orcidid></search><sort><creationdate>2022</creationdate><title>DEEPFAKE Image Synthesis for Data Augmentation</title><author>Waqas, Nawaf ; Safie, Sairul Izwan ; Kadir, Kushsairy Abdul ; Khan, Sheroz ; Kaka Khel, Muhammad Haris</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c408t-f725ac32b6e75d605a57f8beb6094a805003507b06ce76da52ecf2a993ef30cc3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Biomedical imaging</topic><topic>Convolution</topic><topic>Convolutional neural networks</topic><topic>Data augmentation</topic><topic>Data models</topic><topic>Datasets</topic><topic>Deception</topic><topic>Deep learning</topic><topic>DEEPFAKE</topic><topic>Deepfakes</topic><topic>Generative adversarial networks</topic><topic>Image processing</topic><topic>Image resolution</topic><topic>Image segmentation</topic><topic>Medical imaging</topic><topic>PGGAN</topic><topic>self-attention layer</topic><topic>spectral normalization</topic><topic>Synthesis</topic><topic>Training</topic><topic>Training data</topic><topic>unbalanced dataset</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Waqas, Nawaf</creatorcontrib><creatorcontrib>Safie, Sairul Izwan</creatorcontrib><creatorcontrib>Kadir, Kushsairy Abdul</creatorcontrib><creatorcontrib>Khan, Sheroz</creatorcontrib><creatorcontrib>Kaka Khel, Muhammad Haris</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE Open Access Journals</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>IEEE access</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Waqas, Nawaf</au><au>Safie, Sairul Izwan</au><au>Kadir, Kushsairy Abdul</au><au>Khan, Sheroz</au><au>Kaka Khel, Muhammad Haris</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>DEEPFAKE Image Synthesis for Data Augmentation</atitle><jtitle>IEEE access</jtitle><stitle>Access</stitle><date>2022</date><risdate>2022</risdate><volume>10</volume><spage>80847</spage><epage>80857</epage><pages>80847-80857</pages><issn>2169-3536</issn><eissn>2169-3536</eissn><coden>IAECCG</coden><abstract>Field of medical imaging is scarce in terms of a dataset that is reliable and extensive enough to train distinct supervised deep learning models. One way to tackle this problem is to use a Generative Adversarial Network to synthesize DEEPFAKE images to augment the data. DEEPFAKE refers to the transfer of important features from the source image (or video) to the target image (or video), such that the target modality appears to animate the source almost close to reality. In the past decade, medical image processing has made significant advances using the latest state-of-art-methods of deep learning techniques. Supervised deep learning models produce super-human results with the help of huge amount of dataset in a variety of medical image processing and deep learning applications. DEEPFAKE images can be a useful in various applications like translating to different useful and sometimes malicious modalities, unbalanced datasets or increasing the amount of datasets. In this paper the data scarcity has been addressed by using Progressive Growing Generative Adversarial Networks (PGGAN). However, PGGAN consists of convolution layer that suffers from the training-related issues. PGGAN requires a large number of convolution layers in order to obtain high-resolution image training, which makes training a difficult task. In this work, a subjective self-attention layer has been added before &lt;inline-formula&gt; &lt;tex-math notation="LaTeX"&gt;256 \times 256 &lt;/tex-math&gt;&lt;/inline-formula&gt; convolution layer for efficient feature learning and the use of spectral normalization in the discriminator and pixel normalization in the generator for training stabilization - the two tasks resulting into what is referred to as Enhanced-GAN. The performance of Enhanced-GAN is compared to PGGAN performance using the parameters of AM Score and Mode Score. In addition, the strength of Enhanced-GAN and PGGAN synthesized data is evaluated using the U-net supervised deep learning model for segmentation tasks. Dice Coefficient metrics show that U-net trained on Enhanced-GAN DEEPFAKE data optimized with real data performs better than PGGAN DEEPFAKE data with real data.</abstract><cop>Piscataway</cop><pub>IEEE</pub><doi>10.1109/ACCESS.2022.3193668</doi><tpages>11</tpages><orcidid>https://orcid.org/0000-0001-5587-4766</orcidid><orcidid>https://orcid.org/0000-0002-5749-8538</orcidid><orcidid>https://orcid.org/0000-0003-1079-5014</orcidid><orcidid>https://orcid.org/0000-0002-5121-3044</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 2169-3536
ispartof IEEE access, 2022, Vol.10, p.80847-80857
issn 2169-3536
2169-3536
language eng
recordid cdi_proquest_journals_2700412434
source IEEE Open Access Journals; DOAJ Directory of Open Access Journals; Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals
subjects Biomedical imaging
Convolution
Convolutional neural networks
Data augmentation
Data models
Datasets
Deception
Deep learning
DEEPFAKE
Deepfakes
Generative adversarial networks
Image processing
Image resolution
Image segmentation
Medical imaging
PGGAN
self-attention layer
spectral normalization
Synthesis
Training
Training data
unbalanced dataset
title DEEPFAKE Image Synthesis for Data Augmentation
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-02T15%3A02%3A07IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_doaj_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=DEEPFAKE%20Image%20Synthesis%20for%20Data%20Augmentation&rft.jtitle=IEEE%20access&rft.au=Waqas,%20Nawaf&rft.date=2022&rft.volume=10&rft.spage=80847&rft.epage=80857&rft.pages=80847-80857&rft.issn=2169-3536&rft.eissn=2169-3536&rft.coden=IAECCG&rft_id=info:doi/10.1109/ACCESS.2022.3193668&rft_dat=%3Cproquest_doaj_%3E2700412434%3C/proquest_doaj_%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2700412434&rft_id=info:pmid/&rft_ieee_id=9839427&rft_doaj_id=oai_doaj_org_article_368243818ee641de85ce183a2bf84e65&rfr_iscdi=true