Learn while Unlearn: An Iterative Unlearning Framework for Generative Language Models

Recent advancements in machine learning, particularly in Natural Language Processing (NLP), have led to the development of sophisticated models trained on extensive datasets, yet raising concerns about the potential leakage of sensitive information. In response, regulatory measures such as the Europ...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Tang, Haoyu, Liu, Ye, Liu, Xukai, Zhang, Kai, Zhang, Yanghai, Liu, Qi, Chen, Enhong
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Tang, Haoyu
Liu, Ye
Liu, Xukai
Zhang, Kai
Zhang, Yanghai
Liu, Qi
Chen, Enhong
description Recent advancements in machine learning, particularly in Natural Language Processing (NLP), have led to the development of sophisticated models trained on extensive datasets, yet raising concerns about the potential leakage of sensitive information. In response, regulatory measures such as the European Union's General Data Protection Regulation (GDPR) have driven increasing interest in Machine Unlearning techniques, which enable models to selectively forget specific data entries. Early approaches primarily relied on pre-processing methods, while more recent research has shifted towards training-based unlearning techniques. Despite their effectiveness, most existing methods require access to the original training data, which is often inaccessible. Additionally, directly applying unlearning techniques bear the cost of undermining the model's expressive capabilities. To address these challenges, we introduce the Iterative Contrastive Unlearning (ICU) framework, which consists of three core components: A Knowledge Unlearning Induction module designed to remove specific knowledge through an unlearning loss; A Contrastive Learning Enhancement module to preserve the model's expressive capabilities against the pure unlearning goal; And an Iterative Unlearning Refinement module that dynamically assess the unlearning extent on specific data pieces and make iterative update. Experimental results demonstrate the efficacy of our ICU method in unlearning sensitive information while maintaining the model's overall performance, offering a promising solution for privacy-conscious machine learning applications.
doi_str_mv 10.48550/arxiv.2407.20271
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2407_20271</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2407_20271</sourcerecordid><originalsourceid>FETCH-arxiv_primary_2407_202713</originalsourceid><addsrcrecordid>eNpjYJA0NNAzsTA1NdBPLKrILNMzMjEw1zMyMDI35GQI9UlNLMpTKM_IzElVCM3LAfGsFBzzFDxLUosSSzLL4KKZeekKbkWJuanl-UXZCmn5RQruqXkwNT6JeemliempCr75Kak5xTwMrGmJOcWpvFCam0HezTXE2UMX7ID4gqLM3MSiyniQQ-LBDjEmrAIAI6Q-JA</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Learn while Unlearn: An Iterative Unlearning Framework for Generative Language Models</title><source>arXiv.org</source><creator>Tang, Haoyu ; Liu, Ye ; Liu, Xukai ; Zhang, Kai ; Zhang, Yanghai ; Liu, Qi ; Chen, Enhong</creator><creatorcontrib>Tang, Haoyu ; Liu, Ye ; Liu, Xukai ; Zhang, Kai ; Zhang, Yanghai ; Liu, Qi ; Chen, Enhong</creatorcontrib><description>Recent advancements in machine learning, particularly in Natural Language Processing (NLP), have led to the development of sophisticated models trained on extensive datasets, yet raising concerns about the potential leakage of sensitive information. In response, regulatory measures such as the European Union's General Data Protection Regulation (GDPR) have driven increasing interest in Machine Unlearning techniques, which enable models to selectively forget specific data entries. Early approaches primarily relied on pre-processing methods, while more recent research has shifted towards training-based unlearning techniques. Despite their effectiveness, most existing methods require access to the original training data, which is often inaccessible. Additionally, directly applying unlearning techniques bear the cost of undermining the model's expressive capabilities. To address these challenges, we introduce the Iterative Contrastive Unlearning (ICU) framework, which consists of three core components: A Knowledge Unlearning Induction module designed to remove specific knowledge through an unlearning loss; A Contrastive Learning Enhancement module to preserve the model's expressive capabilities against the pure unlearning goal; And an Iterative Unlearning Refinement module that dynamically assess the unlearning extent on specific data pieces and make iterative update. Experimental results demonstrate the efficacy of our ICU method in unlearning sensitive information while maintaining the model's overall performance, offering a promising solution for privacy-conscious machine learning applications.</description><identifier>DOI: 10.48550/arxiv.2407.20271</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Computation and Language ; Computer Science - Learning</subject><creationdate>2024-07</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2407.20271$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2407.20271$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Tang, Haoyu</creatorcontrib><creatorcontrib>Liu, Ye</creatorcontrib><creatorcontrib>Liu, Xukai</creatorcontrib><creatorcontrib>Zhang, Kai</creatorcontrib><creatorcontrib>Zhang, Yanghai</creatorcontrib><creatorcontrib>Liu, Qi</creatorcontrib><creatorcontrib>Chen, Enhong</creatorcontrib><title>Learn while Unlearn: An Iterative Unlearning Framework for Generative Language Models</title><description>Recent advancements in machine learning, particularly in Natural Language Processing (NLP), have led to the development of sophisticated models trained on extensive datasets, yet raising concerns about the potential leakage of sensitive information. In response, regulatory measures such as the European Union's General Data Protection Regulation (GDPR) have driven increasing interest in Machine Unlearning techniques, which enable models to selectively forget specific data entries. Early approaches primarily relied on pre-processing methods, while more recent research has shifted towards training-based unlearning techniques. Despite their effectiveness, most existing methods require access to the original training data, which is often inaccessible. Additionally, directly applying unlearning techniques bear the cost of undermining the model's expressive capabilities. To address these challenges, we introduce the Iterative Contrastive Unlearning (ICU) framework, which consists of three core components: A Knowledge Unlearning Induction module designed to remove specific knowledge through an unlearning loss; A Contrastive Learning Enhancement module to preserve the model's expressive capabilities against the pure unlearning goal; And an Iterative Unlearning Refinement module that dynamically assess the unlearning extent on specific data pieces and make iterative update. Experimental results demonstrate the efficacy of our ICU method in unlearning sensitive information while maintaining the model's overall performance, offering a promising solution for privacy-conscious machine learning applications.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Computation and Language</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNpjYJA0NNAzsTA1NdBPLKrILNMzMjEw1zMyMDI35GQI9UlNLMpTKM_IzElVCM3LAfGsFBzzFDxLUosSSzLL4KKZeekKbkWJuanl-UXZCmn5RQruqXkwNT6JeemliempCr75Kak5xTwMrGmJOcWpvFCam0HezTXE2UMX7ID4gqLM3MSiyniQQ-LBDjEmrAIAI6Q-JA</recordid><startdate>20240725</startdate><enddate>20240725</enddate><creator>Tang, Haoyu</creator><creator>Liu, Ye</creator><creator>Liu, Xukai</creator><creator>Zhang, Kai</creator><creator>Zhang, Yanghai</creator><creator>Liu, Qi</creator><creator>Chen, Enhong</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20240725</creationdate><title>Learn while Unlearn: An Iterative Unlearning Framework for Generative Language Models</title><author>Tang, Haoyu ; Liu, Ye ; Liu, Xukai ; Zhang, Kai ; Zhang, Yanghai ; Liu, Qi ; Chen, Enhong</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-arxiv_primary_2407_202713</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Computation and Language</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Tang, Haoyu</creatorcontrib><creatorcontrib>Liu, Ye</creatorcontrib><creatorcontrib>Liu, Xukai</creatorcontrib><creatorcontrib>Zhang, Kai</creatorcontrib><creatorcontrib>Zhang, Yanghai</creatorcontrib><creatorcontrib>Liu, Qi</creatorcontrib><creatorcontrib>Chen, Enhong</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Tang, Haoyu</au><au>Liu, Ye</au><au>Liu, Xukai</au><au>Zhang, Kai</au><au>Zhang, Yanghai</au><au>Liu, Qi</au><au>Chen, Enhong</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Learn while Unlearn: An Iterative Unlearning Framework for Generative Language Models</atitle><date>2024-07-25</date><risdate>2024</risdate><abstract>Recent advancements in machine learning, particularly in Natural Language Processing (NLP), have led to the development of sophisticated models trained on extensive datasets, yet raising concerns about the potential leakage of sensitive information. In response, regulatory measures such as the European Union's General Data Protection Regulation (GDPR) have driven increasing interest in Machine Unlearning techniques, which enable models to selectively forget specific data entries. Early approaches primarily relied on pre-processing methods, while more recent research has shifted towards training-based unlearning techniques. Despite their effectiveness, most existing methods require access to the original training data, which is often inaccessible. Additionally, directly applying unlearning techniques bear the cost of undermining the model's expressive capabilities. To address these challenges, we introduce the Iterative Contrastive Unlearning (ICU) framework, which consists of three core components: A Knowledge Unlearning Induction module designed to remove specific knowledge through an unlearning loss; A Contrastive Learning Enhancement module to preserve the model's expressive capabilities against the pure unlearning goal; And an Iterative Unlearning Refinement module that dynamically assess the unlearning extent on specific data pieces and make iterative update. Experimental results demonstrate the efficacy of our ICU method in unlearning sensitive information while maintaining the model's overall performance, offering a promising solution for privacy-conscious machine learning applications.</abstract><doi>10.48550/arxiv.2407.20271</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2407.20271
ispartof
issn
language eng
recordid cdi_arxiv_primary_2407_20271
source arXiv.org
subjects Computer Science - Artificial Intelligence
Computer Science - Computation and Language
Computer Science - Learning
title Learn while Unlearn: An Iterative Unlearning Framework for Generative Language Models
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-06T10%3A37%3A36IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Learn%20while%20Unlearn:%20An%20Iterative%20Unlearning%20Framework%20for%20Generative%20Language%20Models&rft.au=Tang,%20Haoyu&rft.date=2024-07-25&rft_id=info:doi/10.48550/arxiv.2407.20271&rft_dat=%3Carxiv_GOX%3E2407_20271%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true