Efficiently Learning at Test-Time: Active Fine-Tuning of LLMs
Recent efforts in fine-tuning language models often rely on automatic data selection, commonly using Nearest Neighbors retrieval from large datasets. However, we theoretically show that this approach tends to select redundant data, limiting its effectiveness or even hurting performance. To address t...
Gespeichert in:
Hauptverfasser: | , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Recent efforts in fine-tuning language models often rely on automatic data
selection, commonly using Nearest Neighbors retrieval from large datasets.
However, we theoretically show that this approach tends to select redundant
data, limiting its effectiveness or even hurting performance. To address this,
we introduce SIFT, a data selection algorithm designed to reduce uncertainty
about the model's response given a prompt, which unifies ideas from retrieval
and active learning. Whereas Nearest Neighbor retrieval typically fails in the
presence of information duplication, SIFT accounts for information duplication
and optimizes the overall information gain of the selected examples. We focus
our evaluations on fine-tuning at test-time for prompt-specific language
modeling on the Pile dataset, and show that SIFT consistently outperforms
Nearest Neighbor retrieval, with minimal computational overhead. Moreover, we
show that our uncertainty estimates can predict the performance gain of
test-time fine-tuning, and use this to develop an adaptive algorithm that
invests test-time compute proportional to realized performance gains. We
provide the $\texttt{activeft}$ (Active Fine-Tuning) library which can be used
as a drop-in replacement for Nearest Neighbor retrieval. |
---|---|
DOI: | 10.48550/arxiv.2410.08020 |