Rethinking Low-Rank Adaptation in Vision: Exploring Head-Level Responsiveness across Diverse Tasks
Low-rank adaptation (LoRA) has shifted the paradigm of adapting pre-trained Vision Transformers (ViT), achieving great efficiency by updating only a subset of tailored parameters to approximate weight updates. However, the multi-head design of the self-attention mechanism, with the heads working in...
Gespeichert in:
Hauptverfasser: | , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Low-rank adaptation (LoRA) has shifted the paradigm of adapting pre-trained
Vision Transformers (ViT), achieving great efficiency by updating only a subset
of tailored parameters to approximate weight updates. However, the multi-head
design of the self-attention mechanism, with the heads working in parallel in
the computation flow, exhibiting similar visual patterns and requiring update
over all of them, incurs unnecessary storage and computational overhead. In
this paper, we propose Head-level responsiveness tuning for low-rank adaptation
(Heart-LoRA). The proposed method explores redundancy among the heads and
selectively activates task-responsive heads, thus enabling fine-grained
head-level tuning. Additionally, given the different responsiveness of heads to
diverse visual tasks, our proposed method dynamically activates a subset of the
approximated heads that are tailored to the current task. Experimental results
show that Heart-LoRA yields superior performance over state-of-the-art PETL
approaches on visual adaptation benchmark datasets. |
---|---|
DOI: | 10.48550/arxiv.2404.08894 |