Limits to scalable evaluation at the frontier: LLM as Judge won't beat twice the data

High quality annotations are increasingly a bottleneck in the explosively growing machine learning ecosystem. Scalable evaluation methods that avoid costly annotation have therefore become an important research ambition. Many hope to use strong existing models in lieu of costly labels to provide che...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Dorner, Florian E, Nastl, Vivian Y, Hardt, Moritz
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Dorner, Florian E
Nastl, Vivian Y
Hardt, Moritz
description High quality annotations are increasingly a bottleneck in the explosively growing machine learning ecosystem. Scalable evaluation methods that avoid costly annotation have therefore become an important research ambition. Many hope to use strong existing models in lieu of costly labels to provide cheap model evaluations. Unfortunately, this method of using models as judges introduces biases, such as self-preferencing, that can distort model comparisons. An emerging family of debiasing tools promises to fix these issues by using a few high quality labels to debias a large number of model judgments. In this paper, we study how far such debiasing methods, in principle, can go. Our main result shows that when the judge is no more accurate than the evaluated model, no debiasing method can decrease the required amount of ground truth labels by more than half. Our result speaks to the severe limitations of the LLM-as-a-judge paradigm at the evaluation frontier where the goal is to assess newly released models that are possibly better than the judge. Through an empirical evaluation, we demonstrate that the sample size savings achievable in practice are even more modest than what our theoretical limit suggests. Along the way, our work provides new observations about debiasing methods for model evaluation, and points out promising avenues for future work.
doi_str_mv 10.48550/arxiv.2410.13341
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2410_13341</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2410_13341</sourcerecordid><originalsourceid>FETCH-arxiv_primary_2410_133413</originalsourceid><addsrcrecordid>eNqFzrEOgkAQRdFtLIz6AVZOZyWCQGJsjcYY7LQmAww6ycKa3QH07xVib_WSl1scpeaB70XbOPbXaF_cepvoewRhGAVjdUu4YnEgBlyOGjNNQC3qBoVNDSggD4LSmlqY7A6S5ALo4NwUd4LO1EuBjPqq45yGtkDBqRqVqB3NfjtRi-Phuj-tBkD6tFyhfac9JB0g4f_iA8yvPSk</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Limits to scalable evaluation at the frontier: LLM as Judge won't beat twice the data</title><source>arXiv.org</source><creator>Dorner, Florian E ; Nastl, Vivian Y ; Hardt, Moritz</creator><creatorcontrib>Dorner, Florian E ; Nastl, Vivian Y ; Hardt, Moritz</creatorcontrib><description>High quality annotations are increasingly a bottleneck in the explosively growing machine learning ecosystem. Scalable evaluation methods that avoid costly annotation have therefore become an important research ambition. Many hope to use strong existing models in lieu of costly labels to provide cheap model evaluations. Unfortunately, this method of using models as judges introduces biases, such as self-preferencing, that can distort model comparisons. An emerging family of debiasing tools promises to fix these issues by using a few high quality labels to debias a large number of model judgments. In this paper, we study how far such debiasing methods, in principle, can go. Our main result shows that when the judge is no more accurate than the evaluated model, no debiasing method can decrease the required amount of ground truth labels by more than half. Our result speaks to the severe limitations of the LLM-as-a-judge paradigm at the evaluation frontier where the goal is to assess newly released models that are possibly better than the judge. Through an empirical evaluation, we demonstrate that the sample size savings achievable in practice are even more modest than what our theoretical limit suggests. Along the way, our work provides new observations about debiasing methods for model evaluation, and points out promising avenues for future work.</description><identifier>DOI: 10.48550/arxiv.2410.13341</identifier><language>eng</language><subject>Computer Science - Learning ; Statistics - Machine Learning</subject><creationdate>2024-10</creationdate><rights>http://creativecommons.org/licenses/by-nc-nd/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2410.13341$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2410.13341$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Dorner, Florian E</creatorcontrib><creatorcontrib>Nastl, Vivian Y</creatorcontrib><creatorcontrib>Hardt, Moritz</creatorcontrib><title>Limits to scalable evaluation at the frontier: LLM as Judge won't beat twice the data</title><description>High quality annotations are increasingly a bottleneck in the explosively growing machine learning ecosystem. Scalable evaluation methods that avoid costly annotation have therefore become an important research ambition. Many hope to use strong existing models in lieu of costly labels to provide cheap model evaluations. Unfortunately, this method of using models as judges introduces biases, such as self-preferencing, that can distort model comparisons. An emerging family of debiasing tools promises to fix these issues by using a few high quality labels to debias a large number of model judgments. In this paper, we study how far such debiasing methods, in principle, can go. Our main result shows that when the judge is no more accurate than the evaluated model, no debiasing method can decrease the required amount of ground truth labels by more than half. Our result speaks to the severe limitations of the LLM-as-a-judge paradigm at the evaluation frontier where the goal is to assess newly released models that are possibly better than the judge. Through an empirical evaluation, we demonstrate that the sample size savings achievable in practice are even more modest than what our theoretical limit suggests. Along the way, our work provides new observations about debiasing methods for model evaluation, and points out promising avenues for future work.</description><subject>Computer Science - Learning</subject><subject>Statistics - Machine Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNqFzrEOgkAQRdFtLIz6AVZOZyWCQGJsjcYY7LQmAww6ycKa3QH07xVib_WSl1scpeaB70XbOPbXaF_cepvoewRhGAVjdUu4YnEgBlyOGjNNQC3qBoVNDSggD4LSmlqY7A6S5ALo4NwUd4LO1EuBjPqq45yGtkDBqRqVqB3NfjtRi-Phuj-tBkD6tFyhfac9JB0g4f_iA8yvPSk</recordid><startdate>20241017</startdate><enddate>20241017</enddate><creator>Dorner, Florian E</creator><creator>Nastl, Vivian Y</creator><creator>Hardt, Moritz</creator><scope>AKY</scope><scope>EPD</scope><scope>GOX</scope></search><sort><creationdate>20241017</creationdate><title>Limits to scalable evaluation at the frontier: LLM as Judge won't beat twice the data</title><author>Dorner, Florian E ; Nastl, Vivian Y ; Hardt, Moritz</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-arxiv_primary_2410_133413</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Learning</topic><topic>Statistics - Machine Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Dorner, Florian E</creatorcontrib><creatorcontrib>Nastl, Vivian Y</creatorcontrib><creatorcontrib>Hardt, Moritz</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv Statistics</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Dorner, Florian E</au><au>Nastl, Vivian Y</au><au>Hardt, Moritz</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Limits to scalable evaluation at the frontier: LLM as Judge won't beat twice the data</atitle><date>2024-10-17</date><risdate>2024</risdate><abstract>High quality annotations are increasingly a bottleneck in the explosively growing machine learning ecosystem. Scalable evaluation methods that avoid costly annotation have therefore become an important research ambition. Many hope to use strong existing models in lieu of costly labels to provide cheap model evaluations. Unfortunately, this method of using models as judges introduces biases, such as self-preferencing, that can distort model comparisons. An emerging family of debiasing tools promises to fix these issues by using a few high quality labels to debias a large number of model judgments. In this paper, we study how far such debiasing methods, in principle, can go. Our main result shows that when the judge is no more accurate than the evaluated model, no debiasing method can decrease the required amount of ground truth labels by more than half. Our result speaks to the severe limitations of the LLM-as-a-judge paradigm at the evaluation frontier where the goal is to assess newly released models that are possibly better than the judge. Through an empirical evaluation, we demonstrate that the sample size savings achievable in practice are even more modest than what our theoretical limit suggests. Along the way, our work provides new observations about debiasing methods for model evaluation, and points out promising avenues for future work.</abstract><doi>10.48550/arxiv.2410.13341</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2410.13341
ispartof
issn
language eng
recordid cdi_arxiv_primary_2410_13341
source arXiv.org
subjects Computer Science - Learning
Statistics - Machine Learning
title Limits to scalable evaluation at the frontier: LLM as Judge won't beat twice the data
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-06T23%3A21%3A28IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Limits%20to%20scalable%20evaluation%20at%20the%20frontier:%20LLM%20as%20Judge%20won't%20beat%20twice%20the%20data&rft.au=Dorner,%20Florian%20E&rft.date=2024-10-17&rft_id=info:doi/10.48550/arxiv.2410.13341&rft_dat=%3Carxiv_GOX%3E2410_13341%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true