Efficiently Learning at Test-Time: Active Fine-Tuning of LLMs

Recent efforts in fine-tuning language models often rely on automatic data selection, commonly using Nearest Neighbors retrieval from large datasets. However, we theoretically show that this approach tends to select redundant data, limiting its effectiveness or even hurting performance. To address t...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Hübotter, Jonas, Bongni, Sascha, Hakimi, Ido, Krause, Andreas
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Hübotter, Jonas
Bongni, Sascha
Hakimi, Ido
Krause, Andreas
description Recent efforts in fine-tuning language models often rely on automatic data selection, commonly using Nearest Neighbors retrieval from large datasets. However, we theoretically show that this approach tends to select redundant data, limiting its effectiveness or even hurting performance. To address this, we introduce SIFT, a data selection algorithm designed to reduce uncertainty about the model's response given a prompt, which unifies ideas from retrieval and active learning. Whereas Nearest Neighbor retrieval typically fails in the presence of information duplication, SIFT accounts for information duplication and optimizes the overall information gain of the selected examples. We focus our evaluations on fine-tuning at test-time for prompt-specific language modeling on the Pile dataset, and show that SIFT consistently outperforms Nearest Neighbor retrieval, with minimal computational overhead. Moreover, we show that our uncertainty estimates can predict the performance gain of test-time fine-tuning, and use this to develop an adaptive algorithm that invests test-time compute proportional to realized performance gains. We provide the $\texttt{activeft}$ (Active Fine-Tuning) library which can be used as a drop-in replacement for Nearest Neighbor retrieval.
doi_str_mv 10.48550/arxiv.2410.08020
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2410_08020</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2410_08020</sourcerecordid><originalsourceid>FETCH-arxiv_primary_2410_080203</originalsourceid><addsrcrecordid>eNpjYJA0NNAzsTA1NdBPLKrILNMzMgEKGFgYGBlwMti6pqVlJmem5pXkVCr4pCYW5WXmpSskliiEpBaX6IZk5qZaKTgml2SWpSq4Zeal6oaUghXkpyn4-PgW8zCwpiXmFKfyQmluBnk31xBnD12wRfEFRZm5iUWV8SAL48EWGhNWAQC8MzQ-</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Efficiently Learning at Test-Time: Active Fine-Tuning of LLMs</title><source>arXiv.org</source><creator>Hübotter, Jonas ; Bongni, Sascha ; Hakimi, Ido ; Krause, Andreas</creator><creatorcontrib>Hübotter, Jonas ; Bongni, Sascha ; Hakimi, Ido ; Krause, Andreas</creatorcontrib><description>Recent efforts in fine-tuning language models often rely on automatic data selection, commonly using Nearest Neighbors retrieval from large datasets. However, we theoretically show that this approach tends to select redundant data, limiting its effectiveness or even hurting performance. To address this, we introduce SIFT, a data selection algorithm designed to reduce uncertainty about the model's response given a prompt, which unifies ideas from retrieval and active learning. Whereas Nearest Neighbor retrieval typically fails in the presence of information duplication, SIFT accounts for information duplication and optimizes the overall information gain of the selected examples. We focus our evaluations on fine-tuning at test-time for prompt-specific language modeling on the Pile dataset, and show that SIFT consistently outperforms Nearest Neighbor retrieval, with minimal computational overhead. Moreover, we show that our uncertainty estimates can predict the performance gain of test-time fine-tuning, and use this to develop an adaptive algorithm that invests test-time compute proportional to realized performance gains. We provide the $\texttt{activeft}$ (Active Fine-Tuning) library which can be used as a drop-in replacement for Nearest Neighbor retrieval.</description><identifier>DOI: 10.48550/arxiv.2410.08020</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Learning</subject><creationdate>2024-10</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2410.08020$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2410.08020$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Hübotter, Jonas</creatorcontrib><creatorcontrib>Bongni, Sascha</creatorcontrib><creatorcontrib>Hakimi, Ido</creatorcontrib><creatorcontrib>Krause, Andreas</creatorcontrib><title>Efficiently Learning at Test-Time: Active Fine-Tuning of LLMs</title><description>Recent efforts in fine-tuning language models often rely on automatic data selection, commonly using Nearest Neighbors retrieval from large datasets. However, we theoretically show that this approach tends to select redundant data, limiting its effectiveness or even hurting performance. To address this, we introduce SIFT, a data selection algorithm designed to reduce uncertainty about the model's response given a prompt, which unifies ideas from retrieval and active learning. Whereas Nearest Neighbor retrieval typically fails in the presence of information duplication, SIFT accounts for information duplication and optimizes the overall information gain of the selected examples. We focus our evaluations on fine-tuning at test-time for prompt-specific language modeling on the Pile dataset, and show that SIFT consistently outperforms Nearest Neighbor retrieval, with minimal computational overhead. Moreover, we show that our uncertainty estimates can predict the performance gain of test-time fine-tuning, and use this to develop an adaptive algorithm that invests test-time compute proportional to realized performance gains. We provide the $\texttt{activeft}$ (Active Fine-Tuning) library which can be used as a drop-in replacement for Nearest Neighbor retrieval.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNpjYJA0NNAzsTA1NdBPLKrILNMzMgEKGFgYGBlwMti6pqVlJmem5pXkVCr4pCYW5WXmpSskliiEpBaX6IZk5qZaKTgml2SWpSq4Zeal6oaUghXkpyn4-PgW8zCwpiXmFKfyQmluBnk31xBnD12wRfEFRZm5iUWV8SAL48EWGhNWAQC8MzQ-</recordid><startdate>20241010</startdate><enddate>20241010</enddate><creator>Hübotter, Jonas</creator><creator>Bongni, Sascha</creator><creator>Hakimi, Ido</creator><creator>Krause, Andreas</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20241010</creationdate><title>Efficiently Learning at Test-Time: Active Fine-Tuning of LLMs</title><author>Hübotter, Jonas ; Bongni, Sascha ; Hakimi, Ido ; Krause, Andreas</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-arxiv_primary_2410_080203</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Hübotter, Jonas</creatorcontrib><creatorcontrib>Bongni, Sascha</creatorcontrib><creatorcontrib>Hakimi, Ido</creatorcontrib><creatorcontrib>Krause, Andreas</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Hübotter, Jonas</au><au>Bongni, Sascha</au><au>Hakimi, Ido</au><au>Krause, Andreas</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Efficiently Learning at Test-Time: Active Fine-Tuning of LLMs</atitle><date>2024-10-10</date><risdate>2024</risdate><abstract>Recent efforts in fine-tuning language models often rely on automatic data selection, commonly using Nearest Neighbors retrieval from large datasets. However, we theoretically show that this approach tends to select redundant data, limiting its effectiveness or even hurting performance. To address this, we introduce SIFT, a data selection algorithm designed to reduce uncertainty about the model's response given a prompt, which unifies ideas from retrieval and active learning. Whereas Nearest Neighbor retrieval typically fails in the presence of information duplication, SIFT accounts for information duplication and optimizes the overall information gain of the selected examples. We focus our evaluations on fine-tuning at test-time for prompt-specific language modeling on the Pile dataset, and show that SIFT consistently outperforms Nearest Neighbor retrieval, with minimal computational overhead. Moreover, we show that our uncertainty estimates can predict the performance gain of test-time fine-tuning, and use this to develop an adaptive algorithm that invests test-time compute proportional to realized performance gains. We provide the $\texttt{activeft}$ (Active Fine-Tuning) library which can be used as a drop-in replacement for Nearest Neighbor retrieval.</abstract><doi>10.48550/arxiv.2410.08020</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2410.08020
ispartof
issn
language eng
recordid cdi_arxiv_primary_2410_08020
source arXiv.org
subjects Computer Science - Artificial Intelligence
Computer Science - Learning
title Efficiently Learning at Test-Time: Active Fine-Tuning of LLMs
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-01T03%3A22%3A25IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Efficiently%20Learning%20at%20Test-Time:%20Active%20Fine-Tuning%20of%20LLMs&rft.au=H%C3%BCbotter,%20Jonas&rft.date=2024-10-10&rft_id=info:doi/10.48550/arxiv.2410.08020&rft_dat=%3Carxiv_GOX%3E2410_08020%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true