ShareLoRA: Parameter Efficient and Robust Large Language Model Fine-tuning via Shared Low-Rank Adaptation
This study introduces an approach to optimize Parameter Efficient Fine Tuning (PEFT) for Pretrained Language Models (PLMs) by implementing a Shared Low Rank Adaptation (ShareLoRA). By strategically deploying ShareLoRA across different layers and adapting it for the Query, Key, and Value components o...
Gespeichert in:
Hauptverfasser: | , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | This study introduces an approach to optimize Parameter Efficient Fine Tuning
(PEFT) for Pretrained Language Models (PLMs) by implementing a Shared Low Rank
Adaptation (ShareLoRA). By strategically deploying ShareLoRA across different
layers and adapting it for the Query, Key, and Value components of
self-attention layers, we achieve a substantial reduction in the number of
training parameters and memory usage. Importantly, ShareLoRA not only maintains
model performance but also exhibits robustness in both classification and
generation tasks across a variety of models, including RoBERTa, GPT-2, LLaMA
and LLaMA2. It demonstrates superior transfer learning capabilities compared to
standard LoRA applications and mitigates overfitting by sharing weights across
layers. Our findings affirm that ShareLoRA effectively boosts parameter
efficiency while ensuring scalable and high-quality performance across
different language model architectures. |
---|---|
DOI: | 10.48550/arxiv.2406.10785 |