Unharmful Backdoor-based Client-side Watermarking in Federated Learning
Protecting intellectual property (IP) in federated learning (FL) is increasingly important as clients contribute proprietary data to collaboratively train models. Model watermarking, particularly through backdoor-based methods, has emerged as a popular approach for verifying ownership and contributi...
Gespeichert in:
Hauptverfasser: | , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Protecting intellectual property (IP) in federated learning (FL) is
increasingly important as clients contribute proprietary data to
collaboratively train models. Model watermarking, particularly through
backdoor-based methods, has emerged as a popular approach for verifying
ownership and contributions in deep neural networks trained via FL. By
manipulating their datasets, clients can embed a secret pattern, resulting in
non-intuitive predictions that serve as proof of participation, useful for
claiming incentives or IP co-ownership. However, this technique faces practical
challenges: client watermarks can collide, leading to ambiguous ownership
claims, and malicious clients may exploit watermarks to inject harmful
backdoors, jeopardizing model integrity. To address these issues, we propose
Sanitizer, a server-side method that ensures client-embedded backdoors cannot
be triggered on natural queries in harmful ways. It identifies subnets within
client-submitted models, extracts backdoors throughout the FL process, and
confines them to harmless, client-specific input subspaces. This approach not
only enhances Sanitizer's efficiency but also resolves conflicts when clients
use similar triggers with different target labels. Our empirical results
demonstrate that Sanitizer achieves near-perfect success in verifying client
contributions while mitigating the risks of malicious watermark use.
Additionally, it reduces GPU memory consumption by 85% and cuts processing time
by at least 5 times compared to the baseline. |
---|---|
DOI: | 10.48550/arxiv.2410.21179 |