Explanation-based data-free model extraction attacks

Deep learning (DL) has dramatically pushed the previous limits of various tasks, ranging from computer vision to natural language processing. Despite its success, the lack of model explanations thwarts the usage of these techniques in life-critical domains, e.g., medical diagnosis and self-driving s...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:World wide web (Bussum) 2023-09, Vol.26 (5), p.3081-3092
Hauptverfasser: Yan, Anli, Hou, Ruitao, Yan, Hongyang, Liu, Xiaozhang
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 3092
container_issue 5
container_start_page 3081
container_title World wide web (Bussum)
container_volume 26
creator Yan, Anli
Hou, Ruitao
Yan, Hongyang
Liu, Xiaozhang
description Deep learning (DL) has dramatically pushed the previous limits of various tasks, ranging from computer vision to natural language processing. Despite its success, the lack of model explanations thwarts the usage of these techniques in life-critical domains, e.g., medical diagnosis and self-driving systems. To date, the core technology to solve the explainable issue is explainable artificial intelligence (XAI). XAI methods have been developed to produce human-understandable explanations by leveraging intermediate results of the DL models, e.g., gradients and model parameters. While the effectiveness of XAI methods has been demonstrated in benign environments, their privacy against model extraction attacks (i.e., attacks at the model confidentially) requires to be studied. To this end, this paper proposes DMEAE, a d ata-free m odel e xtraction a ttack using e xplanation-guided, to explore XAI privacy threats. Compared with previous works, DMEAE does not require collecting any data and utilizes model explanation loss. Specifically, DMEAE creates synthetic data using a generative model with model explanation loss items. Extensive evaluations verify the effectiveness and efficiency of the proposed attack strategy on SVHN and CIFAR-10 datasets. We hope that our research can provide insights for the development of practical tools to trade off the relationship between privacy and model explanations.
doi_str_mv 10.1007/s11280-023-01150-6
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2875642294</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2875642294</sourcerecordid><originalsourceid>FETCH-LOGICAL-c319t-d0dbcf436a3a1b919f7a99df990723235f76dcc98da0475f9a3364d68b2fbea33</originalsourceid><addsrcrecordid>eNp9kE1LAzEQhoMoWKt_wNOC5-jkY5PNUUq1QsGLgrcwu0mktd2tSQr135u6gjdP8w487ww8hFwzuGUA-i4xxhugwAUFxmqg6oRMWK0FZZKJ05JFo0qu387JRUprAFDCsAmR88Nugz3m1dDTFpN3lcOMNETvq-3g_KbyhxyxOwIV5ozdR7okZwE3yV_9zil5fZi_zBZ0-fz4NLtf0k4wk6kD13ZBCoUCWWuYCRqNccEY0FxwUQetXNeZxiFIXQeDQijpVNPy0PqyTMnNeHcXh8-9T9muh33sy0vLG10rybmRheIj1cUhpeiD3cXVFuOXZWCPduxoxxY79seOVaUkxlIqcP_u49_pf1rfZgVnEg</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2875642294</pqid></control><display><type>article</type><title>Explanation-based data-free model extraction attacks</title><source>SpringerLink Journals</source><creator>Yan, Anli ; Hou, Ruitao ; Yan, Hongyang ; Liu, Xiaozhang</creator><creatorcontrib>Yan, Anli ; Hou, Ruitao ; Yan, Hongyang ; Liu, Xiaozhang</creatorcontrib><description>Deep learning (DL) has dramatically pushed the previous limits of various tasks, ranging from computer vision to natural language processing. Despite its success, the lack of model explanations thwarts the usage of these techniques in life-critical domains, e.g., medical diagnosis and self-driving systems. To date, the core technology to solve the explainable issue is explainable artificial intelligence (XAI). XAI methods have been developed to produce human-understandable explanations by leveraging intermediate results of the DL models, e.g., gradients and model parameters. While the effectiveness of XAI methods has been demonstrated in benign environments, their privacy against model extraction attacks (i.e., attacks at the model confidentially) requires to be studied. To this end, this paper proposes DMEAE, a d ata-free m odel e xtraction a ttack using e xplanation-guided, to explore XAI privacy threats. Compared with previous works, DMEAE does not require collecting any data and utilizes model explanation loss. Specifically, DMEAE creates synthetic data using a generative model with model explanation loss items. Extensive evaluations verify the effectiveness and efficiency of the proposed attack strategy on SVHN and CIFAR-10 datasets. We hope that our research can provide insights for the development of practical tools to trade off the relationship between privacy and model explanations.</description><identifier>ISSN: 1386-145X</identifier><identifier>EISSN: 1573-1413</identifier><identifier>DOI: 10.1007/s11280-023-01150-6</identifier><language>eng</language><publisher>New York: Springer US</publisher><subject>Computer Science ; Computer vision ; Database Management ; Effectiveness ; Explainable artificial intelligence ; Information Systems Applications (incl.Internet) ; Natural language processing ; Operating Systems ; Privacy ; Special Issue on Privacy and Security in Machine Learning ; Synthetic data</subject><ispartof>World wide web (Bussum), 2023-09, Vol.26 (5), p.3081-3092</ispartof><rights>The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2023. Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c319t-d0dbcf436a3a1b919f7a99df990723235f76dcc98da0475f9a3364d68b2fbea33</citedby><cites>FETCH-LOGICAL-c319t-d0dbcf436a3a1b919f7a99df990723235f76dcc98da0475f9a3364d68b2fbea33</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s11280-023-01150-6$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s11280-023-01150-6$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>314,780,784,27924,27925,41488,42557,51319</link.rule.ids></links><search><creatorcontrib>Yan, Anli</creatorcontrib><creatorcontrib>Hou, Ruitao</creatorcontrib><creatorcontrib>Yan, Hongyang</creatorcontrib><creatorcontrib>Liu, Xiaozhang</creatorcontrib><title>Explanation-based data-free model extraction attacks</title><title>World wide web (Bussum)</title><addtitle>World Wide Web</addtitle><description>Deep learning (DL) has dramatically pushed the previous limits of various tasks, ranging from computer vision to natural language processing. Despite its success, the lack of model explanations thwarts the usage of these techniques in life-critical domains, e.g., medical diagnosis and self-driving systems. To date, the core technology to solve the explainable issue is explainable artificial intelligence (XAI). XAI methods have been developed to produce human-understandable explanations by leveraging intermediate results of the DL models, e.g., gradients and model parameters. While the effectiveness of XAI methods has been demonstrated in benign environments, their privacy against model extraction attacks (i.e., attacks at the model confidentially) requires to be studied. To this end, this paper proposes DMEAE, a d ata-free m odel e xtraction a ttack using e xplanation-guided, to explore XAI privacy threats. Compared with previous works, DMEAE does not require collecting any data and utilizes model explanation loss. Specifically, DMEAE creates synthetic data using a generative model with model explanation loss items. Extensive evaluations verify the effectiveness and efficiency of the proposed attack strategy on SVHN and CIFAR-10 datasets. We hope that our research can provide insights for the development of practical tools to trade off the relationship between privacy and model explanations.</description><subject>Computer Science</subject><subject>Computer vision</subject><subject>Database Management</subject><subject>Effectiveness</subject><subject>Explainable artificial intelligence</subject><subject>Information Systems Applications (incl.Internet)</subject><subject>Natural language processing</subject><subject>Operating Systems</subject><subject>Privacy</subject><subject>Special Issue on Privacy and Security in Machine Learning</subject><subject>Synthetic data</subject><issn>1386-145X</issn><issn>1573-1413</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><sourceid>GNUQQ</sourceid><recordid>eNp9kE1LAzEQhoMoWKt_wNOC5-jkY5PNUUq1QsGLgrcwu0mktd2tSQr135u6gjdP8w487ww8hFwzuGUA-i4xxhugwAUFxmqg6oRMWK0FZZKJ05JFo0qu387JRUprAFDCsAmR88Nugz3m1dDTFpN3lcOMNETvq-3g_KbyhxyxOwIV5ozdR7okZwE3yV_9zil5fZi_zBZ0-fz4NLtf0k4wk6kD13ZBCoUCWWuYCRqNccEY0FxwUQetXNeZxiFIXQeDQijpVNPy0PqyTMnNeHcXh8-9T9muh33sy0vLG10rybmRheIj1cUhpeiD3cXVFuOXZWCPduxoxxY79seOVaUkxlIqcP_u49_pf1rfZgVnEg</recordid><startdate>20230901</startdate><enddate>20230901</enddate><creator>Yan, Anli</creator><creator>Hou, Ruitao</creator><creator>Yan, Hongyang</creator><creator>Liu, Xiaozhang</creator><general>Springer US</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7SC</scope><scope>7XB</scope><scope>8AL</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>8FK</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>GNUQQ</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>K7-</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>M0N</scope><scope>P5Z</scope><scope>P62</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>Q9U</scope></search><sort><creationdate>20230901</creationdate><title>Explanation-based data-free model extraction attacks</title><author>Yan, Anli ; Hou, Ruitao ; Yan, Hongyang ; Liu, Xiaozhang</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c319t-d0dbcf436a3a1b919f7a99df990723235f76dcc98da0475f9a3364d68b2fbea33</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science</topic><topic>Computer vision</topic><topic>Database Management</topic><topic>Effectiveness</topic><topic>Explainable artificial intelligence</topic><topic>Information Systems Applications (incl.Internet)</topic><topic>Natural language processing</topic><topic>Operating Systems</topic><topic>Privacy</topic><topic>Special Issue on Privacy and Security in Machine Learning</topic><topic>Synthetic data</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Yan, Anli</creatorcontrib><creatorcontrib>Hou, Ruitao</creatorcontrib><creatorcontrib>Yan, Hongyang</creatorcontrib><creatorcontrib>Liu, Xiaozhang</creatorcontrib><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Computer and Information Systems Abstracts</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>Computing Database (Alumni Edition)</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>Advanced Technologies &amp; Aerospace Collection</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>ProQuest Central Student</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Computer Science Collection</collection><collection>Computer Science Database</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>Computing Database</collection><collection>Advanced Technologies &amp; Aerospace Database</collection><collection>ProQuest Advanced Technologies &amp; Aerospace Collection</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central Basic</collection><jtitle>World wide web (Bussum)</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Yan, Anli</au><au>Hou, Ruitao</au><au>Yan, Hongyang</au><au>Liu, Xiaozhang</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Explanation-based data-free model extraction attacks</atitle><jtitle>World wide web (Bussum)</jtitle><stitle>World Wide Web</stitle><date>2023-09-01</date><risdate>2023</risdate><volume>26</volume><issue>5</issue><spage>3081</spage><epage>3092</epage><pages>3081-3092</pages><issn>1386-145X</issn><eissn>1573-1413</eissn><abstract>Deep learning (DL) has dramatically pushed the previous limits of various tasks, ranging from computer vision to natural language processing. Despite its success, the lack of model explanations thwarts the usage of these techniques in life-critical domains, e.g., medical diagnosis and self-driving systems. To date, the core technology to solve the explainable issue is explainable artificial intelligence (XAI). XAI methods have been developed to produce human-understandable explanations by leveraging intermediate results of the DL models, e.g., gradients and model parameters. While the effectiveness of XAI methods has been demonstrated in benign environments, their privacy against model extraction attacks (i.e., attacks at the model confidentially) requires to be studied. To this end, this paper proposes DMEAE, a d ata-free m odel e xtraction a ttack using e xplanation-guided, to explore XAI privacy threats. Compared with previous works, DMEAE does not require collecting any data and utilizes model explanation loss. Specifically, DMEAE creates synthetic data using a generative model with model explanation loss items. Extensive evaluations verify the effectiveness and efficiency of the proposed attack strategy on SVHN and CIFAR-10 datasets. We hope that our research can provide insights for the development of practical tools to trade off the relationship between privacy and model explanations.</abstract><cop>New York</cop><pub>Springer US</pub><doi>10.1007/s11280-023-01150-6</doi><tpages>12</tpages></addata></record>
fulltext fulltext
identifier ISSN: 1386-145X
ispartof World wide web (Bussum), 2023-09, Vol.26 (5), p.3081-3092
issn 1386-145X
1573-1413
language eng
recordid cdi_proquest_journals_2875642294
source SpringerLink Journals
subjects Computer Science
Computer vision
Database Management
Effectiveness
Explainable artificial intelligence
Information Systems Applications (incl.Internet)
Natural language processing
Operating Systems
Privacy
Special Issue on Privacy and Security in Machine Learning
Synthetic data
title Explanation-based data-free model extraction attacks
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-27T12%3A19%3A44IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Explanation-based%20data-free%20model%20extraction%20attacks&rft.jtitle=World%20wide%20web%20(Bussum)&rft.au=Yan,%20Anli&rft.date=2023-09-01&rft.volume=26&rft.issue=5&rft.spage=3081&rft.epage=3092&rft.pages=3081-3092&rft.issn=1386-145X&rft.eissn=1573-1413&rft_id=info:doi/10.1007/s11280-023-01150-6&rft_dat=%3Cproquest_cross%3E2875642294%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2875642294&rft_id=info:pmid/&rfr_iscdi=true