On lower bounds for the bias-variance trade-off

It is a common phenomenon that for high-dimensional and nonparametric statistical models, rate-optimal estimators balance squared bias and variance. Although this balancing is widely observed, little is known whether methods exist that could avoid the trade-off between bias and variance. We propose...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:The Annals of statistics 2023-08, Vol.51 (4), p.1510
Hauptverfasser: Derumigny, Alexis, Schmidt-Hieber, Johannes
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue 4
container_start_page 1510
container_title The Annals of statistics
container_volume 51
creator Derumigny, Alexis
Schmidt-Hieber, Johannes
description It is a common phenomenon that for high-dimensional and nonparametric statistical models, rate-optimal estimators balance squared bias and variance. Although this balancing is widely observed, little is known whether methods exist that could avoid the trade-off between bias and variance. We propose a general strategy to obtain lower bounds on the variance of any estimator with bias smaller than a prespecified bound. This shows to which extent the bias-variance trade-off is unavoidable and allows to quantify the loss of performance for methods that do not obey it. The approach is based on a number of abstract lower bounds for the variance involving the change of expectation with respect to different probability measures as well as information measures such as the Kullback–Leibler or χ2 -divergence. Some of these inequalities rely on a new concept of information matrices. In a second part of the article, the abstract lower bounds are applied to several statistical models including the Gaussian white noise model, a boundary estimation problem, the Gaussian sequence model and the high-dimensional linear regression model. For these specific statistical applications, different types of bias-variance trade-offs occur that vary considerably in their strength. For the trade-off between integrated squared bias and integrated variance in the Gaussian white noise model, we propose to combine the general strategy for lower bounds with a reduction technique. This allows us to reduce the original problem to a lower bound on the bias-variance trade-off for estimators with additional symmetry properties in a simpler statistical model. In the Gaussian sequence model, different phase transitions of the bias-variance trade-off occur. Although there is a non-trivial interplay between bias and variance, the rate of the squared bias and the variance do not have to be balanced in order to achieve the minimax estimation rate.
doi_str_mv 10.1214/23-AOS2279
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2882570777</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2882570777</sourcerecordid><originalsourceid>FETCH-LOGICAL-c174t-21c6c2b039aa47b279f13843b94eb00587fca9f9ceb3a579345f49abbeca3aaf3</originalsourceid><addsrcrecordid>eNotkMFKAzEURYMoOFY3fkHAnRCb5GUmk2UpWoXCLNR1eEkTnFInNZkq_r0j7epuDvdcLiG3gj8IKdRcAlt0r1Jqc0YqKZqWtaZpzknFueGshkZdkqtStpzz2iioyLwb6C79hExdOgybQmPKdPwI1PVY2DfmHgcf6JhxE1iK8ZpcRNyVcHPKGXl_enxbPrN1t3pZLtbMC61GJoVvvHQcDKLSbtoTBbQKnFHBTe5WR48mGh8cYK0NqDoqg84Fj4AYYUbujr37nL4OoYx2mw55mJRWtq2sNddaT9T9kfI5lZJDtPvcf2L-tYLb_0OsBHs6BP4AKf1Rkw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2882570777</pqid></control><display><type>article</type><title>On lower bounds for the bias-variance trade-off</title><source>Project Euclid Complete</source><creator>Derumigny, Alexis ; Schmidt-Hieber, Johannes</creator><creatorcontrib>Derumigny, Alexis ; Schmidt-Hieber, Johannes</creatorcontrib><description>It is a common phenomenon that for high-dimensional and nonparametric statistical models, rate-optimal estimators balance squared bias and variance. Although this balancing is widely observed, little is known whether methods exist that could avoid the trade-off between bias and variance. We propose a general strategy to obtain lower bounds on the variance of any estimator with bias smaller than a prespecified bound. This shows to which extent the bias-variance trade-off is unavoidable and allows to quantify the loss of performance for methods that do not obey it. The approach is based on a number of abstract lower bounds for the variance involving the change of expectation with respect to different probability measures as well as information measures such as the Kullback–Leibler or χ2 -divergence. Some of these inequalities rely on a new concept of information matrices. In a second part of the article, the abstract lower bounds are applied to several statistical models including the Gaussian white noise model, a boundary estimation problem, the Gaussian sequence model and the high-dimensional linear regression model. For these specific statistical applications, different types of bias-variance trade-offs occur that vary considerably in their strength. For the trade-off between integrated squared bias and integrated variance in the Gaussian white noise model, we propose to combine the general strategy for lower bounds with a reduction technique. This allows us to reduce the original problem to a lower bound on the bias-variance trade-off for estimators with additional symmetry properties in a simpler statistical model. In the Gaussian sequence model, different phase transitions of the bias-variance trade-off occur. Although there is a non-trivial interplay between bias and variance, the rate of the squared bias and the variance do not have to be balanced in order to achieve the minimax estimation rate.</description><identifier>ISSN: 0090-5364</identifier><identifier>EISSN: 2168-8966</identifier><identifier>DOI: 10.1214/23-AOS2279</identifier><language>eng</language><publisher>Hayward: Institute of Mathematical Statistics</publisher><subject>Bias ; Divergence ; Estimating techniques ; Estimators ; Lower bounds ; Minimax technique ; Normal distribution ; Phase transitions ; Regression models ; Statistical analysis ; Statistical models ; Tradeoffs ; Variance ; Variance analysis ; White noise</subject><ispartof>The Annals of statistics, 2023-08, Vol.51 (4), p.1510</ispartof><rights>Copyright Institute of Mathematical Statistics Aug 2023</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c174t-21c6c2b039aa47b279f13843b94eb00587fca9f9ceb3a579345f49abbeca3aaf3</citedby><cites>FETCH-LOGICAL-c174t-21c6c2b039aa47b279f13843b94eb00587fca9f9ceb3a579345f49abbeca3aaf3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,27924,27925</link.rule.ids></links><search><creatorcontrib>Derumigny, Alexis</creatorcontrib><creatorcontrib>Schmidt-Hieber, Johannes</creatorcontrib><title>On lower bounds for the bias-variance trade-off</title><title>The Annals of statistics</title><description>It is a common phenomenon that for high-dimensional and nonparametric statistical models, rate-optimal estimators balance squared bias and variance. Although this balancing is widely observed, little is known whether methods exist that could avoid the trade-off between bias and variance. We propose a general strategy to obtain lower bounds on the variance of any estimator with bias smaller than a prespecified bound. This shows to which extent the bias-variance trade-off is unavoidable and allows to quantify the loss of performance for methods that do not obey it. The approach is based on a number of abstract lower bounds for the variance involving the change of expectation with respect to different probability measures as well as information measures such as the Kullback–Leibler or χ2 -divergence. Some of these inequalities rely on a new concept of information matrices. In a second part of the article, the abstract lower bounds are applied to several statistical models including the Gaussian white noise model, a boundary estimation problem, the Gaussian sequence model and the high-dimensional linear regression model. For these specific statistical applications, different types of bias-variance trade-offs occur that vary considerably in their strength. For the trade-off between integrated squared bias and integrated variance in the Gaussian white noise model, we propose to combine the general strategy for lower bounds with a reduction technique. This allows us to reduce the original problem to a lower bound on the bias-variance trade-off for estimators with additional symmetry properties in a simpler statistical model. In the Gaussian sequence model, different phase transitions of the bias-variance trade-off occur. Although there is a non-trivial interplay between bias and variance, the rate of the squared bias and the variance do not have to be balanced in order to achieve the minimax estimation rate.</description><subject>Bias</subject><subject>Divergence</subject><subject>Estimating techniques</subject><subject>Estimators</subject><subject>Lower bounds</subject><subject>Minimax technique</subject><subject>Normal distribution</subject><subject>Phase transitions</subject><subject>Regression models</subject><subject>Statistical analysis</subject><subject>Statistical models</subject><subject>Tradeoffs</subject><subject>Variance</subject><subject>Variance analysis</subject><subject>White noise</subject><issn>0090-5364</issn><issn>2168-8966</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><recordid>eNotkMFKAzEURYMoOFY3fkHAnRCb5GUmk2UpWoXCLNR1eEkTnFInNZkq_r0j7epuDvdcLiG3gj8IKdRcAlt0r1Jqc0YqKZqWtaZpzknFueGshkZdkqtStpzz2iioyLwb6C79hExdOgybQmPKdPwI1PVY2DfmHgcf6JhxE1iK8ZpcRNyVcHPKGXl_enxbPrN1t3pZLtbMC61GJoVvvHQcDKLSbtoTBbQKnFHBTe5WR48mGh8cYK0NqDoqg84Fj4AYYUbujr37nL4OoYx2mw55mJRWtq2sNddaT9T9kfI5lZJDtPvcf2L-tYLb_0OsBHs6BP4AKf1Rkw</recordid><startdate>20230801</startdate><enddate>20230801</enddate><creator>Derumigny, Alexis</creator><creator>Schmidt-Hieber, Johannes</creator><general>Institute of Mathematical Statistics</general><scope>AAYXX</scope><scope>CITATION</scope><scope>JQ2</scope></search><sort><creationdate>20230801</creationdate><title>On lower bounds for the bias-variance trade-off</title><author>Derumigny, Alexis ; Schmidt-Hieber, Johannes</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c174t-21c6c2b039aa47b279f13843b94eb00587fca9f9ceb3a579345f49abbeca3aaf3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Bias</topic><topic>Divergence</topic><topic>Estimating techniques</topic><topic>Estimators</topic><topic>Lower bounds</topic><topic>Minimax technique</topic><topic>Normal distribution</topic><topic>Phase transitions</topic><topic>Regression models</topic><topic>Statistical analysis</topic><topic>Statistical models</topic><topic>Tradeoffs</topic><topic>Variance</topic><topic>Variance analysis</topic><topic>White noise</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Derumigny, Alexis</creatorcontrib><creatorcontrib>Schmidt-Hieber, Johannes</creatorcontrib><collection>CrossRef</collection><collection>ProQuest Computer Science Collection</collection><jtitle>The Annals of statistics</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Derumigny, Alexis</au><au>Schmidt-Hieber, Johannes</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>On lower bounds for the bias-variance trade-off</atitle><jtitle>The Annals of statistics</jtitle><date>2023-08-01</date><risdate>2023</risdate><volume>51</volume><issue>4</issue><spage>1510</spage><pages>1510-</pages><issn>0090-5364</issn><eissn>2168-8966</eissn><abstract>It is a common phenomenon that for high-dimensional and nonparametric statistical models, rate-optimal estimators balance squared bias and variance. Although this balancing is widely observed, little is known whether methods exist that could avoid the trade-off between bias and variance. We propose a general strategy to obtain lower bounds on the variance of any estimator with bias smaller than a prespecified bound. This shows to which extent the bias-variance trade-off is unavoidable and allows to quantify the loss of performance for methods that do not obey it. The approach is based on a number of abstract lower bounds for the variance involving the change of expectation with respect to different probability measures as well as information measures such as the Kullback–Leibler or χ2 -divergence. Some of these inequalities rely on a new concept of information matrices. In a second part of the article, the abstract lower bounds are applied to several statistical models including the Gaussian white noise model, a boundary estimation problem, the Gaussian sequence model and the high-dimensional linear regression model. For these specific statistical applications, different types of bias-variance trade-offs occur that vary considerably in their strength. For the trade-off between integrated squared bias and integrated variance in the Gaussian white noise model, we propose to combine the general strategy for lower bounds with a reduction technique. This allows us to reduce the original problem to a lower bound on the bias-variance trade-off for estimators with additional symmetry properties in a simpler statistical model. In the Gaussian sequence model, different phase transitions of the bias-variance trade-off occur. Although there is a non-trivial interplay between bias and variance, the rate of the squared bias and the variance do not have to be balanced in order to achieve the minimax estimation rate.</abstract><cop>Hayward</cop><pub>Institute of Mathematical Statistics</pub><doi>10.1214/23-AOS2279</doi></addata></record>
fulltext fulltext
identifier ISSN: 0090-5364
ispartof The Annals of statistics, 2023-08, Vol.51 (4), p.1510
issn 0090-5364
2168-8966
language eng
recordid cdi_proquest_journals_2882570777
source Project Euclid Complete
subjects Bias
Divergence
Estimating techniques
Estimators
Lower bounds
Minimax technique
Normal distribution
Phase transitions
Regression models
Statistical analysis
Statistical models
Tradeoffs
Variance
Variance analysis
White noise
title On lower bounds for the bias-variance trade-off
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-28T14%3A25%3A33IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=On%20lower%20bounds%20for%20the%20bias-variance%20trade-off&rft.jtitle=The%20Annals%20of%20statistics&rft.au=Derumigny,%20Alexis&rft.date=2023-08-01&rft.volume=51&rft.issue=4&rft.spage=1510&rft.pages=1510-&rft.issn=0090-5364&rft.eissn=2168-8966&rft_id=info:doi/10.1214/23-AOS2279&rft_dat=%3Cproquest_cross%3E2882570777%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2882570777&rft_id=info:pmid/&rfr_iscdi=true