Proof of Unlearning: Definitions and Instantiation
The "Right to be Forgotten" rule in machine learning (ML) practice enables some individual data to be deleted from a trained model, as pursued by recently developed machine unlearning techniques. To truly comply with the rule, a natural and necessary step is to verify if the individual dat...
Gespeichert in:
Hauptverfasser: | , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | The "Right to be Forgotten" rule in machine learning (ML) practice enables
some individual data to be deleted from a trained model, as pursued by recently
developed machine unlearning techniques. To truly comply with the rule, a
natural and necessary step is to verify if the individual data are indeed
deleted after unlearning. Yet, previous parameter-space verification metrics
may be easily evaded by a distrustful model trainer. Thus, Thudi et al.
recently present a call to action on algorithm-level verification in USENIX
Security'22.
We respond to the call, by reconsidering the unlearning problem in the
scenario of machine learning as a service (MLaaS), and proposing a new
definition framework for Proof of Unlearning (PoUL) on algorithm level.
Specifically, our PoUL definitions (i) enforce correctness properties on both
the pre and post phases of unlearning, so as to prevent the state-of-the-art
forging attacks; (ii) highlight proper practicality requirements of both the
prover and verifier sides with minimal invasiveness to the off-the-shelf
service pipeline and computational workloads. Under the definition framework,
we subsequently present a trusted hardware-empowered instantiation using SGX
enclave, by logically incorporating an authentication layer for tracing the
data lineage with a proving layer for supporting the audit of learning. We
customize authenticated data structures to support large out-of-enclave storage
with simple operation logic, and meanwhile, enable proving complex unlearning
logic with affordable memory footprints in the enclave. We finally validate the
feasibility of the proposed instantiation with a proof-of-concept
implementation and multi-dimensional performance evaluation. |
---|---|
DOI: | 10.48550/arxiv.2210.11334 |