FedPFT: Federated Proxy Fine-Tuning of Foundation Models
Adapting Foundation Models (FMs) for downstream tasks through Federated Learning (FL) emerges a promising strategy for protecting data privacy and valuable FMs. Existing methods fine-tune FM by allocating sub-FM to clients in FL, however, leading to suboptimal performance due to insufficient tuning...
Gespeichert in:
Hauptverfasser: | , , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Adapting Foundation Models (FMs) for downstream tasks through Federated
Learning (FL) emerges a promising strategy for protecting data privacy and
valuable FMs. Existing methods fine-tune FM by allocating sub-FM to clients in
FL, however, leading to suboptimal performance due to insufficient tuning and
inevitable error accumulations of gradients. In this paper, we propose
Federated Proxy Fine-Tuning (FedPFT), a novel method enhancing FMs adaptation
in downstream tasks through FL by two key modules. First, the sub-FM
construction module employs a layer-wise compression approach, facilitating
comprehensive FM fine-tuning across all layers by emphasizing those crucial
neurons. Second, the sub-FM alignment module conducts a two-step
distillations-layer-level and neuron-level-before and during FL fine-tuning
respectively, to reduce error of gradient by accurately aligning sub-FM with FM
under theoretical guarantees. Experimental results on seven commonly used
datasets (i.e., four text and three vision) demonstrate the superiority of
FedPFT. |
---|---|
DOI: | 10.48550/arxiv.2404.11536 |