On lower bounds for the bias-variance trade-off
It is a common phenomenon that for high-dimensional and nonparametric statistical models, rate-optimal estimators balance squared bias and variance. Although this balancing is widely observed, little is known whether methods exist that could avoid the trade-off between bias and variance. We propose...
Gespeichert in:
Veröffentlicht in: | The Annals of statistics 2023-08, Vol.51 (4), p.1510 |
---|---|
Hauptverfasser: | , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | 4 |
container_start_page | 1510 |
container_title | The Annals of statistics |
container_volume | 51 |
creator | Derumigny, Alexis Schmidt-Hieber, Johannes |
description | It is a common phenomenon that for high-dimensional and nonparametric statistical models, rate-optimal estimators balance squared bias and variance. Although this balancing is widely observed, little is known whether methods exist that could avoid the trade-off between bias and variance. We propose a general strategy to obtain lower bounds on the variance of any estimator with bias smaller than a prespecified bound. This shows to which extent the bias-variance trade-off is unavoidable and allows to quantify the loss of performance for methods that do not obey it. The approach is based on a number of abstract lower bounds for the variance involving the change of expectation with respect to different probability measures as well as information measures such as the Kullback–Leibler or χ2 -divergence. Some of these inequalities rely on a new concept of information matrices. In a second part of the article, the abstract lower bounds are applied to several statistical models including the Gaussian white noise model, a boundary estimation problem, the Gaussian sequence model and the high-dimensional linear regression model. For these specific statistical applications, different types of bias-variance trade-offs occur that vary considerably in their strength. For the trade-off between integrated squared bias and integrated variance in the Gaussian white noise model, we propose to combine the general strategy for lower bounds with a reduction technique. This allows us to reduce the original problem to a lower bound on the bias-variance trade-off for estimators with additional symmetry properties in a simpler statistical model. In the Gaussian sequence model, different phase transitions of the bias-variance trade-off occur. Although there is a non-trivial interplay between bias and variance, the rate of the squared bias and the variance do not have to be balanced in order to achieve the minimax estimation rate. |
doi_str_mv | 10.1214/23-AOS2279 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2882570777</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2882570777</sourcerecordid><originalsourceid>FETCH-LOGICAL-c174t-21c6c2b039aa47b279f13843b94eb00587fca9f9ceb3a579345f49abbeca3aaf3</originalsourceid><addsrcrecordid>eNotkMFKAzEURYMoOFY3fkHAnRCb5GUmk2UpWoXCLNR1eEkTnFInNZkq_r0j7epuDvdcLiG3gj8IKdRcAlt0r1Jqc0YqKZqWtaZpzknFueGshkZdkqtStpzz2iioyLwb6C79hExdOgybQmPKdPwI1PVY2DfmHgcf6JhxE1iK8ZpcRNyVcHPKGXl_enxbPrN1t3pZLtbMC61GJoVvvHQcDKLSbtoTBbQKnFHBTe5WR48mGh8cYK0NqDoqg84Fj4AYYUbujr37nL4OoYx2mw55mJRWtq2sNddaT9T9kfI5lZJDtPvcf2L-tYLb_0OsBHs6BP4AKf1Rkw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2882570777</pqid></control><display><type>article</type><title>On lower bounds for the bias-variance trade-off</title><source>Project Euclid Complete</source><creator>Derumigny, Alexis ; Schmidt-Hieber, Johannes</creator><creatorcontrib>Derumigny, Alexis ; Schmidt-Hieber, Johannes</creatorcontrib><description>It is a common phenomenon that for high-dimensional and nonparametric statistical models, rate-optimal estimators balance squared bias and variance. Although this balancing is widely observed, little is known whether methods exist that could avoid the trade-off between bias and variance. We propose a general strategy to obtain lower bounds on the variance of any estimator with bias smaller than a prespecified bound. This shows to which extent the bias-variance trade-off is unavoidable and allows to quantify the loss of performance for methods that do not obey it. The approach is based on a number of abstract lower bounds for the variance involving the change of expectation with respect to different probability measures as well as information measures such as the Kullback–Leibler or χ2 -divergence. Some of these inequalities rely on a new concept of information matrices. In a second part of the article, the abstract lower bounds are applied to several statistical models including the Gaussian white noise model, a boundary estimation problem, the Gaussian sequence model and the high-dimensional linear regression model. For these specific statistical applications, different types of bias-variance trade-offs occur that vary considerably in their strength. For the trade-off between integrated squared bias and integrated variance in the Gaussian white noise model, we propose to combine the general strategy for lower bounds with a reduction technique. This allows us to reduce the original problem to a lower bound on the bias-variance trade-off for estimators with additional symmetry properties in a simpler statistical model. In the Gaussian sequence model, different phase transitions of the bias-variance trade-off occur. Although there is a non-trivial interplay between bias and variance, the rate of the squared bias and the variance do not have to be balanced in order to achieve the minimax estimation rate.</description><identifier>ISSN: 0090-5364</identifier><identifier>EISSN: 2168-8966</identifier><identifier>DOI: 10.1214/23-AOS2279</identifier><language>eng</language><publisher>Hayward: Institute of Mathematical Statistics</publisher><subject>Bias ; Divergence ; Estimating techniques ; Estimators ; Lower bounds ; Minimax technique ; Normal distribution ; Phase transitions ; Regression models ; Statistical analysis ; Statistical models ; Tradeoffs ; Variance ; Variance analysis ; White noise</subject><ispartof>The Annals of statistics, 2023-08, Vol.51 (4), p.1510</ispartof><rights>Copyright Institute of Mathematical Statistics Aug 2023</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c174t-21c6c2b039aa47b279f13843b94eb00587fca9f9ceb3a579345f49abbeca3aaf3</citedby><cites>FETCH-LOGICAL-c174t-21c6c2b039aa47b279f13843b94eb00587fca9f9ceb3a579345f49abbeca3aaf3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,27924,27925</link.rule.ids></links><search><creatorcontrib>Derumigny, Alexis</creatorcontrib><creatorcontrib>Schmidt-Hieber, Johannes</creatorcontrib><title>On lower bounds for the bias-variance trade-off</title><title>The Annals of statistics</title><description>It is a common phenomenon that for high-dimensional and nonparametric statistical models, rate-optimal estimators balance squared bias and variance. Although this balancing is widely observed, little is known whether methods exist that could avoid the trade-off between bias and variance. We propose a general strategy to obtain lower bounds on the variance of any estimator with bias smaller than a prespecified bound. This shows to which extent the bias-variance trade-off is unavoidable and allows to quantify the loss of performance for methods that do not obey it. The approach is based on a number of abstract lower bounds for the variance involving the change of expectation with respect to different probability measures as well as information measures such as the Kullback–Leibler or χ2 -divergence. Some of these inequalities rely on a new concept of information matrices. In a second part of the article, the abstract lower bounds are applied to several statistical models including the Gaussian white noise model, a boundary estimation problem, the Gaussian sequence model and the high-dimensional linear regression model. For these specific statistical applications, different types of bias-variance trade-offs occur that vary considerably in their strength. For the trade-off between integrated squared bias and integrated variance in the Gaussian white noise model, we propose to combine the general strategy for lower bounds with a reduction technique. This allows us to reduce the original problem to a lower bound on the bias-variance trade-off for estimators with additional symmetry properties in a simpler statistical model. In the Gaussian sequence model, different phase transitions of the bias-variance trade-off occur. Although there is a non-trivial interplay between bias and variance, the rate of the squared bias and the variance do not have to be balanced in order to achieve the minimax estimation rate.</description><subject>Bias</subject><subject>Divergence</subject><subject>Estimating techniques</subject><subject>Estimators</subject><subject>Lower bounds</subject><subject>Minimax technique</subject><subject>Normal distribution</subject><subject>Phase transitions</subject><subject>Regression models</subject><subject>Statistical analysis</subject><subject>Statistical models</subject><subject>Tradeoffs</subject><subject>Variance</subject><subject>Variance analysis</subject><subject>White noise</subject><issn>0090-5364</issn><issn>2168-8966</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><recordid>eNotkMFKAzEURYMoOFY3fkHAnRCb5GUmk2UpWoXCLNR1eEkTnFInNZkq_r0j7epuDvdcLiG3gj8IKdRcAlt0r1Jqc0YqKZqWtaZpzknFueGshkZdkqtStpzz2iioyLwb6C79hExdOgybQmPKdPwI1PVY2DfmHgcf6JhxE1iK8ZpcRNyVcHPKGXl_enxbPrN1t3pZLtbMC61GJoVvvHQcDKLSbtoTBbQKnFHBTe5WR48mGh8cYK0NqDoqg84Fj4AYYUbujr37nL4OoYx2mw55mJRWtq2sNddaT9T9kfI5lZJDtPvcf2L-tYLb_0OsBHs6BP4AKf1Rkw</recordid><startdate>20230801</startdate><enddate>20230801</enddate><creator>Derumigny, Alexis</creator><creator>Schmidt-Hieber, Johannes</creator><general>Institute of Mathematical Statistics</general><scope>AAYXX</scope><scope>CITATION</scope><scope>JQ2</scope></search><sort><creationdate>20230801</creationdate><title>On lower bounds for the bias-variance trade-off</title><author>Derumigny, Alexis ; Schmidt-Hieber, Johannes</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c174t-21c6c2b039aa47b279f13843b94eb00587fca9f9ceb3a579345f49abbeca3aaf3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Bias</topic><topic>Divergence</topic><topic>Estimating techniques</topic><topic>Estimators</topic><topic>Lower bounds</topic><topic>Minimax technique</topic><topic>Normal distribution</topic><topic>Phase transitions</topic><topic>Regression models</topic><topic>Statistical analysis</topic><topic>Statistical models</topic><topic>Tradeoffs</topic><topic>Variance</topic><topic>Variance analysis</topic><topic>White noise</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Derumigny, Alexis</creatorcontrib><creatorcontrib>Schmidt-Hieber, Johannes</creatorcontrib><collection>CrossRef</collection><collection>ProQuest Computer Science Collection</collection><jtitle>The Annals of statistics</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Derumigny, Alexis</au><au>Schmidt-Hieber, Johannes</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>On lower bounds for the bias-variance trade-off</atitle><jtitle>The Annals of statistics</jtitle><date>2023-08-01</date><risdate>2023</risdate><volume>51</volume><issue>4</issue><spage>1510</spage><pages>1510-</pages><issn>0090-5364</issn><eissn>2168-8966</eissn><abstract>It is a common phenomenon that for high-dimensional and nonparametric statistical models, rate-optimal estimators balance squared bias and variance. Although this balancing is widely observed, little is known whether methods exist that could avoid the trade-off between bias and variance. We propose a general strategy to obtain lower bounds on the variance of any estimator with bias smaller than a prespecified bound. This shows to which extent the bias-variance trade-off is unavoidable and allows to quantify the loss of performance for methods that do not obey it. The approach is based on a number of abstract lower bounds for the variance involving the change of expectation with respect to different probability measures as well as information measures such as the Kullback–Leibler or χ2 -divergence. Some of these inequalities rely on a new concept of information matrices. In a second part of the article, the abstract lower bounds are applied to several statistical models including the Gaussian white noise model, a boundary estimation problem, the Gaussian sequence model and the high-dimensional linear regression model. For these specific statistical applications, different types of bias-variance trade-offs occur that vary considerably in their strength. For the trade-off between integrated squared bias and integrated variance in the Gaussian white noise model, we propose to combine the general strategy for lower bounds with a reduction technique. This allows us to reduce the original problem to a lower bound on the bias-variance trade-off for estimators with additional symmetry properties in a simpler statistical model. In the Gaussian sequence model, different phase transitions of the bias-variance trade-off occur. Although there is a non-trivial interplay between bias and variance, the rate of the squared bias and the variance do not have to be balanced in order to achieve the minimax estimation rate.</abstract><cop>Hayward</cop><pub>Institute of Mathematical Statistics</pub><doi>10.1214/23-AOS2279</doi></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0090-5364 |
ispartof | The Annals of statistics, 2023-08, Vol.51 (4), p.1510 |
issn | 0090-5364 2168-8966 |
language | eng |
recordid | cdi_proquest_journals_2882570777 |
source | Project Euclid Complete |
subjects | Bias Divergence Estimating techniques Estimators Lower bounds Minimax technique Normal distribution Phase transitions Regression models Statistical analysis Statistical models Tradeoffs Variance Variance analysis White noise |
title | On lower bounds for the bias-variance trade-off |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-28T14%3A25%3A33IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=On%20lower%20bounds%20for%20the%20bias-variance%20trade-off&rft.jtitle=The%20Annals%20of%20statistics&rft.au=Derumigny,%20Alexis&rft.date=2023-08-01&rft.volume=51&rft.issue=4&rft.spage=1510&rft.pages=1510-&rft.issn=0090-5364&rft.eissn=2168-8966&rft_id=info:doi/10.1214/23-AOS2279&rft_dat=%3Cproquest_cross%3E2882570777%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2882570777&rft_id=info:pmid/&rfr_iscdi=true |