DEEPFAKE Image Synthesis for Data Augmentation
Field of medical imaging is scarce in terms of a dataset that is reliable and extensive enough to train distinct supervised deep learning models. One way to tackle this problem is to use a Generative Adversarial Network to synthesize DEEPFAKE images to augment the data. DEEPFAKE refers to the transf...
Gespeichert in:
Veröffentlicht in: | IEEE access 2022, Vol.10, p.80847-80857 |
---|---|
Hauptverfasser: | , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 80857 |
---|---|
container_issue | |
container_start_page | 80847 |
container_title | IEEE access |
container_volume | 10 |
creator | Waqas, Nawaf Safie, Sairul Izwan Kadir, Kushsairy Abdul Khan, Sheroz Kaka Khel, Muhammad Haris |
description | Field of medical imaging is scarce in terms of a dataset that is reliable and extensive enough to train distinct supervised deep learning models. One way to tackle this problem is to use a Generative Adversarial Network to synthesize DEEPFAKE images to augment the data. DEEPFAKE refers to the transfer of important features from the source image (or video) to the target image (or video), such that the target modality appears to animate the source almost close to reality. In the past decade, medical image processing has made significant advances using the latest state-of-art-methods of deep learning techniques. Supervised deep learning models produce super-human results with the help of huge amount of dataset in a variety of medical image processing and deep learning applications. DEEPFAKE images can be a useful in various applications like translating to different useful and sometimes malicious modalities, unbalanced datasets or increasing the amount of datasets. In this paper the data scarcity has been addressed by using Progressive Growing Generative Adversarial Networks (PGGAN). However, PGGAN consists of convolution layer that suffers from the training-related issues. PGGAN requires a large number of convolution layers in order to obtain high-resolution image training, which makes training a difficult task. In this work, a subjective self-attention layer has been added before 256 \times 256 convolution layer for efficient feature learning and the use of spectral normalization in the discriminator and pixel normalization in the generator for training stabilization - the two tasks resulting into what is referred to as Enhanced-GAN. The performance of Enhanced-GAN is compared to PGGAN performance using the parameters of AM Score and Mode Score. In addition, the strength of Enhanced-GAN and PGGAN synthesized data is evaluated using the U-net supervised deep learning model for segmentation tasks. Dice Coefficient metrics show that U-net trained on Enhanced-GAN DEEPFAKE data optimized with real data performs better than PGGAN DEEPFAKE data with real data. |
doi_str_mv | 10.1109/ACCESS.2022.3193668 |
format | Article |
fullrecord | <record><control><sourceid>proquest_doaj_</sourceid><recordid>TN_cdi_proquest_journals_2700412434</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9839427</ieee_id><doaj_id>oai_doaj_org_article_368243818ee641de85ce183a2bf84e65</doaj_id><sourcerecordid>2700412434</sourcerecordid><originalsourceid>FETCH-LOGICAL-c408t-f725ac32b6e75d605a57f8beb6094a805003507b06ce76da52ecf2a993ef30cc3</originalsourceid><addsrcrecordid>eNpNkE1Lw0AQhoMoWGp_QS8Bz4n7kf06hjTVYkGhel42m0lNabN1Nz3035uaIs5lhpd53xmeKJpjlGKM1FNeFOVmkxJESEqxopzLm2hCMFcJZZTf_pvvo1kIOzSUHCQmJlG6KMv3Zf5axquD2UK8OXf9F4Q2xI3z8cL0Js5P2wN0velb1z1Ed43ZB5hd-zT6XJYfxUuyfnteFfk6sRmSfdIIwoylpOIgWM0RM0w0soKKI5UZiRhClCFRIW5B8NowArYhRikKDUXW0mm0GnNrZ3b66NuD8WftTKt_Bee32vi-tXvQlEuSUYklAM9wDZJZwJIaUjUyA86GrMcx6-jd9wlCr3fu5LvhfU0EQhke7NmwRcct610IHpq_qxjpC2c9ctYXzvrKeXDNR1cLAH8OJanKiKA_3ot15A</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2700412434</pqid></control><display><type>article</type><title>DEEPFAKE Image Synthesis for Data Augmentation</title><source>IEEE Open Access Journals</source><source>DOAJ Directory of Open Access Journals</source><source>Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals</source><creator>Waqas, Nawaf ; Safie, Sairul Izwan ; Kadir, Kushsairy Abdul ; Khan, Sheroz ; Kaka Khel, Muhammad Haris</creator><creatorcontrib>Waqas, Nawaf ; Safie, Sairul Izwan ; Kadir, Kushsairy Abdul ; Khan, Sheroz ; Kaka Khel, Muhammad Haris</creatorcontrib><description>Field of medical imaging is scarce in terms of a dataset that is reliable and extensive enough to train distinct supervised deep learning models. One way to tackle this problem is to use a Generative Adversarial Network to synthesize DEEPFAKE images to augment the data. DEEPFAKE refers to the transfer of important features from the source image (or video) to the target image (or video), such that the target modality appears to animate the source almost close to reality. In the past decade, medical image processing has made significant advances using the latest state-of-art-methods of deep learning techniques. Supervised deep learning models produce super-human results with the help of huge amount of dataset in a variety of medical image processing and deep learning applications. DEEPFAKE images can be a useful in various applications like translating to different useful and sometimes malicious modalities, unbalanced datasets or increasing the amount of datasets. In this paper the data scarcity has been addressed by using Progressive Growing Generative Adversarial Networks (PGGAN). However, PGGAN consists of convolution layer that suffers from the training-related issues. PGGAN requires a large number of convolution layers in order to obtain high-resolution image training, which makes training a difficult task. In this work, a subjective self-attention layer has been added before <inline-formula> <tex-math notation="LaTeX">256 \times 256 </tex-math></inline-formula> convolution layer for efficient feature learning and the use of spectral normalization in the discriminator and pixel normalization in the generator for training stabilization - the two tasks resulting into what is referred to as Enhanced-GAN. The performance of Enhanced-GAN is compared to PGGAN performance using the parameters of AM Score and Mode Score. In addition, the strength of Enhanced-GAN and PGGAN synthesized data is evaluated using the U-net supervised deep learning model for segmentation tasks. Dice Coefficient metrics show that U-net trained on Enhanced-GAN DEEPFAKE data optimized with real data performs better than PGGAN DEEPFAKE data with real data.</description><identifier>ISSN: 2169-3536</identifier><identifier>EISSN: 2169-3536</identifier><identifier>DOI: 10.1109/ACCESS.2022.3193668</identifier><identifier>CODEN: IAECCG</identifier><language>eng</language><publisher>Piscataway: IEEE</publisher><subject>Biomedical imaging ; Convolution ; Convolutional neural networks ; Data augmentation ; Data models ; Datasets ; Deception ; Deep learning ; DEEPFAKE ; Deepfakes ; Generative adversarial networks ; Image processing ; Image resolution ; Image segmentation ; Medical imaging ; PGGAN ; self-attention layer ; spectral normalization ; Synthesis ; Training ; Training data ; unbalanced dataset</subject><ispartof>IEEE access, 2022, Vol.10, p.80847-80857</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2022</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c408t-f725ac32b6e75d605a57f8beb6094a805003507b06ce76da52ecf2a993ef30cc3</citedby><cites>FETCH-LOGICAL-c408t-f725ac32b6e75d605a57f8beb6094a805003507b06ce76da52ecf2a993ef30cc3</cites><orcidid>0000-0001-5587-4766 ; 0000-0002-5749-8538 ; 0000-0003-1079-5014 ; 0000-0002-5121-3044</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9839427$$EHTML$$P50$$Gieee$$Hfree_for_read</linktohtml><link.rule.ids>314,776,780,860,2096,4010,27610,27900,27901,27902,54908</link.rule.ids></links><search><creatorcontrib>Waqas, Nawaf</creatorcontrib><creatorcontrib>Safie, Sairul Izwan</creatorcontrib><creatorcontrib>Kadir, Kushsairy Abdul</creatorcontrib><creatorcontrib>Khan, Sheroz</creatorcontrib><creatorcontrib>Kaka Khel, Muhammad Haris</creatorcontrib><title>DEEPFAKE Image Synthesis for Data Augmentation</title><title>IEEE access</title><addtitle>Access</addtitle><description>Field of medical imaging is scarce in terms of a dataset that is reliable and extensive enough to train distinct supervised deep learning models. One way to tackle this problem is to use a Generative Adversarial Network to synthesize DEEPFAKE images to augment the data. DEEPFAKE refers to the transfer of important features from the source image (or video) to the target image (or video), such that the target modality appears to animate the source almost close to reality. In the past decade, medical image processing has made significant advances using the latest state-of-art-methods of deep learning techniques. Supervised deep learning models produce super-human results with the help of huge amount of dataset in a variety of medical image processing and deep learning applications. DEEPFAKE images can be a useful in various applications like translating to different useful and sometimes malicious modalities, unbalanced datasets or increasing the amount of datasets. In this paper the data scarcity has been addressed by using Progressive Growing Generative Adversarial Networks (PGGAN). However, PGGAN consists of convolution layer that suffers from the training-related issues. PGGAN requires a large number of convolution layers in order to obtain high-resolution image training, which makes training a difficult task. In this work, a subjective self-attention layer has been added before <inline-formula> <tex-math notation="LaTeX">256 \times 256 </tex-math></inline-formula> convolution layer for efficient feature learning and the use of spectral normalization in the discriminator and pixel normalization in the generator for training stabilization - the two tasks resulting into what is referred to as Enhanced-GAN. The performance of Enhanced-GAN is compared to PGGAN performance using the parameters of AM Score and Mode Score. In addition, the strength of Enhanced-GAN and PGGAN synthesized data is evaluated using the U-net supervised deep learning model for segmentation tasks. Dice Coefficient metrics show that U-net trained on Enhanced-GAN DEEPFAKE data optimized with real data performs better than PGGAN DEEPFAKE data with real data.</description><subject>Biomedical imaging</subject><subject>Convolution</subject><subject>Convolutional neural networks</subject><subject>Data augmentation</subject><subject>Data models</subject><subject>Datasets</subject><subject>Deception</subject><subject>Deep learning</subject><subject>DEEPFAKE</subject><subject>Deepfakes</subject><subject>Generative adversarial networks</subject><subject>Image processing</subject><subject>Image resolution</subject><subject>Image segmentation</subject><subject>Medical imaging</subject><subject>PGGAN</subject><subject>self-attention layer</subject><subject>spectral normalization</subject><subject>Synthesis</subject><subject>Training</subject><subject>Training data</subject><subject>unbalanced dataset</subject><issn>2169-3536</issn><issn>2169-3536</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>ESBDL</sourceid><sourceid>RIE</sourceid><sourceid>DOA</sourceid><recordid>eNpNkE1Lw0AQhoMoWGp_QS8Bz4n7kf06hjTVYkGhel42m0lNabN1Nz3035uaIs5lhpd53xmeKJpjlGKM1FNeFOVmkxJESEqxopzLm2hCMFcJZZTf_pvvo1kIOzSUHCQmJlG6KMv3Zf5axquD2UK8OXf9F4Q2xI3z8cL0Js5P2wN0velb1z1Ed43ZB5hd-zT6XJYfxUuyfnteFfk6sRmSfdIIwoylpOIgWM0RM0w0soKKI5UZiRhClCFRIW5B8NowArYhRikKDUXW0mm0GnNrZ3b66NuD8WftTKt_Bee32vi-tXvQlEuSUYklAM9wDZJZwJIaUjUyA86GrMcx6-jd9wlCr3fu5LvhfU0EQhke7NmwRcct610IHpq_qxjpC2c9ctYXzvrKeXDNR1cLAH8OJanKiKA_3ot15A</recordid><startdate>2022</startdate><enddate>2022</enddate><creator>Waqas, Nawaf</creator><creator>Safie, Sairul Izwan</creator><creator>Kadir, Kushsairy Abdul</creator><creator>Khan, Sheroz</creator><creator>Kaka Khel, Muhammad Haris</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>ESBDL</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>7SR</scope><scope>8BQ</scope><scope>8FD</scope><scope>JG9</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>DOA</scope><orcidid>https://orcid.org/0000-0001-5587-4766</orcidid><orcidid>https://orcid.org/0000-0002-5749-8538</orcidid><orcidid>https://orcid.org/0000-0003-1079-5014</orcidid><orcidid>https://orcid.org/0000-0002-5121-3044</orcidid></search><sort><creationdate>2022</creationdate><title>DEEPFAKE Image Synthesis for Data Augmentation</title><author>Waqas, Nawaf ; Safie, Sairul Izwan ; Kadir, Kushsairy Abdul ; Khan, Sheroz ; Kaka Khel, Muhammad Haris</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c408t-f725ac32b6e75d605a57f8beb6094a805003507b06ce76da52ecf2a993ef30cc3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Biomedical imaging</topic><topic>Convolution</topic><topic>Convolutional neural networks</topic><topic>Data augmentation</topic><topic>Data models</topic><topic>Datasets</topic><topic>Deception</topic><topic>Deep learning</topic><topic>DEEPFAKE</topic><topic>Deepfakes</topic><topic>Generative adversarial networks</topic><topic>Image processing</topic><topic>Image resolution</topic><topic>Image segmentation</topic><topic>Medical imaging</topic><topic>PGGAN</topic><topic>self-attention layer</topic><topic>spectral normalization</topic><topic>Synthesis</topic><topic>Training</topic><topic>Training data</topic><topic>unbalanced dataset</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Waqas, Nawaf</creatorcontrib><creatorcontrib>Safie, Sairul Izwan</creatorcontrib><creatorcontrib>Kadir, Kushsairy Abdul</creatorcontrib><creatorcontrib>Khan, Sheroz</creatorcontrib><creatorcontrib>Kaka Khel, Muhammad Haris</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE Open Access Journals</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics & Communications Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>IEEE access</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Waqas, Nawaf</au><au>Safie, Sairul Izwan</au><au>Kadir, Kushsairy Abdul</au><au>Khan, Sheroz</au><au>Kaka Khel, Muhammad Haris</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>DEEPFAKE Image Synthesis for Data Augmentation</atitle><jtitle>IEEE access</jtitle><stitle>Access</stitle><date>2022</date><risdate>2022</risdate><volume>10</volume><spage>80847</spage><epage>80857</epage><pages>80847-80857</pages><issn>2169-3536</issn><eissn>2169-3536</eissn><coden>IAECCG</coden><abstract>Field of medical imaging is scarce in terms of a dataset that is reliable and extensive enough to train distinct supervised deep learning models. One way to tackle this problem is to use a Generative Adversarial Network to synthesize DEEPFAKE images to augment the data. DEEPFAKE refers to the transfer of important features from the source image (or video) to the target image (or video), such that the target modality appears to animate the source almost close to reality. In the past decade, medical image processing has made significant advances using the latest state-of-art-methods of deep learning techniques. Supervised deep learning models produce super-human results with the help of huge amount of dataset in a variety of medical image processing and deep learning applications. DEEPFAKE images can be a useful in various applications like translating to different useful and sometimes malicious modalities, unbalanced datasets or increasing the amount of datasets. In this paper the data scarcity has been addressed by using Progressive Growing Generative Adversarial Networks (PGGAN). However, PGGAN consists of convolution layer that suffers from the training-related issues. PGGAN requires a large number of convolution layers in order to obtain high-resolution image training, which makes training a difficult task. In this work, a subjective self-attention layer has been added before <inline-formula> <tex-math notation="LaTeX">256 \times 256 </tex-math></inline-formula> convolution layer for efficient feature learning and the use of spectral normalization in the discriminator and pixel normalization in the generator for training stabilization - the two tasks resulting into what is referred to as Enhanced-GAN. The performance of Enhanced-GAN is compared to PGGAN performance using the parameters of AM Score and Mode Score. In addition, the strength of Enhanced-GAN and PGGAN synthesized data is evaluated using the U-net supervised deep learning model for segmentation tasks. Dice Coefficient metrics show that U-net trained on Enhanced-GAN DEEPFAKE data optimized with real data performs better than PGGAN DEEPFAKE data with real data.</abstract><cop>Piscataway</cop><pub>IEEE</pub><doi>10.1109/ACCESS.2022.3193668</doi><tpages>11</tpages><orcidid>https://orcid.org/0000-0001-5587-4766</orcidid><orcidid>https://orcid.org/0000-0002-5749-8538</orcidid><orcidid>https://orcid.org/0000-0003-1079-5014</orcidid><orcidid>https://orcid.org/0000-0002-5121-3044</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 2169-3536 |
ispartof | IEEE access, 2022, Vol.10, p.80847-80857 |
issn | 2169-3536 2169-3536 |
language | eng |
recordid | cdi_proquest_journals_2700412434 |
source | IEEE Open Access Journals; DOAJ Directory of Open Access Journals; Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals |
subjects | Biomedical imaging Convolution Convolutional neural networks Data augmentation Data models Datasets Deception Deep learning DEEPFAKE Deepfakes Generative adversarial networks Image processing Image resolution Image segmentation Medical imaging PGGAN self-attention layer spectral normalization Synthesis Training Training data unbalanced dataset |
title | DEEPFAKE Image Synthesis for Data Augmentation |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-02T15%3A02%3A07IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_doaj_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=DEEPFAKE%20Image%20Synthesis%20for%20Data%20Augmentation&rft.jtitle=IEEE%20access&rft.au=Waqas,%20Nawaf&rft.date=2022&rft.volume=10&rft.spage=80847&rft.epage=80857&rft.pages=80847-80857&rft.issn=2169-3536&rft.eissn=2169-3536&rft.coden=IAECCG&rft_id=info:doi/10.1109/ACCESS.2022.3193668&rft_dat=%3Cproquest_doaj_%3E2700412434%3C/proquest_doaj_%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2700412434&rft_id=info:pmid/&rft_ieee_id=9839427&rft_doaj_id=oai_doaj_org_article_368243818ee641de85ce183a2bf84e65&rfr_iscdi=true |