A Kernel Test of Goodness of Fit

We propose a nonparametric statistical test for goodness-of-fit: given a set of samples, the test determines how likely it is that these were generated from a target density function. The measure of goodness-of-fit is a divergence constructed via Stein's method using functions from a Reproducin...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Chwialkowski, Kacper, Strathmann, Heiko, Gretton, Arthur
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Chwialkowski, Kacper
Strathmann, Heiko
Gretton, Arthur
description We propose a nonparametric statistical test for goodness-of-fit: given a set of samples, the test determines how likely it is that these were generated from a target density function. The measure of goodness-of-fit is a divergence constructed via Stein's method using functions from a Reproducing Kernel Hilbert Space. Our test statistic is based on an empirical estimate of this divergence, taking the form of a V-statistic in terms of the log gradients of the target density and the kernel. We derive a statistical test, both for i.i.d. and non-i.i.d. samples, where we estimate the null distribution quantiles using a wild bootstrap procedure. We apply our test to quantifying convergence of approximate Markov Chain Monte Carlo methods, statistical model criticism, and evaluating quality of fit vs model complexity in nonparametric density estimation.
doi_str_mv 10.48550/arxiv.1602.02964
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_1602_02964</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>1602_02964</sourcerecordid><originalsourceid>FETCH-LOGICAL-a674-dcda59dd49c0f9a62fcba78d75fa55fcdc8006ee71605e7994494f382ccf40833</originalsourceid><addsrcrecordid>eNotzr0OgjAABOAuDgZ9ACf7AmDpf0dCBI0kLuyk9ichQTCUGH17BZ3uprsPgF2KEioZQwc9vtpnknKEE4QVp2sAM3hxY-86WLswwcHDchhs70KYe9FOG7Dyugtu-88I1MWxzk9xdS3PeVbFmgsaW2M1U9ZSZZBXmmNvblpIK5jXjHljjUSIOye-38wJpShV1BOJjfEUSUIisP_NLsTmMbZ3Pb6bmdosVPIBPa03ZA</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>A Kernel Test of Goodness of Fit</title><source>arXiv.org</source><creator>Chwialkowski, Kacper ; Strathmann, Heiko ; Gretton, Arthur</creator><creatorcontrib>Chwialkowski, Kacper ; Strathmann, Heiko ; Gretton, Arthur</creatorcontrib><description>We propose a nonparametric statistical test for goodness-of-fit: given a set of samples, the test determines how likely it is that these were generated from a target density function. The measure of goodness-of-fit is a divergence constructed via Stein's method using functions from a Reproducing Kernel Hilbert Space. Our test statistic is based on an empirical estimate of this divergence, taking the form of a V-statistic in terms of the log gradients of the target density and the kernel. We derive a statistical test, both for i.i.d. and non-i.i.d. samples, where we estimate the null distribution quantiles using a wild bootstrap procedure. We apply our test to quantifying convergence of approximate Markov Chain Monte Carlo methods, statistical model criticism, and evaluating quality of fit vs model complexity in nonparametric density estimation.</description><identifier>DOI: 10.48550/arxiv.1602.02964</identifier><language>eng</language><subject>Statistics - Machine Learning</subject><creationdate>2016-02</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,782,887</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/1602.02964$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.1602.02964$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Chwialkowski, Kacper</creatorcontrib><creatorcontrib>Strathmann, Heiko</creatorcontrib><creatorcontrib>Gretton, Arthur</creatorcontrib><title>A Kernel Test of Goodness of Fit</title><description>We propose a nonparametric statistical test for goodness-of-fit: given a set of samples, the test determines how likely it is that these were generated from a target density function. The measure of goodness-of-fit is a divergence constructed via Stein's method using functions from a Reproducing Kernel Hilbert Space. Our test statistic is based on an empirical estimate of this divergence, taking the form of a V-statistic in terms of the log gradients of the target density and the kernel. We derive a statistical test, both for i.i.d. and non-i.i.d. samples, where we estimate the null distribution quantiles using a wild bootstrap procedure. We apply our test to quantifying convergence of approximate Markov Chain Monte Carlo methods, statistical model criticism, and evaluating quality of fit vs model complexity in nonparametric density estimation.</description><subject>Statistics - Machine Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2016</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotzr0OgjAABOAuDgZ9ACf7AmDpf0dCBI0kLuyk9ichQTCUGH17BZ3uprsPgF2KEioZQwc9vtpnknKEE4QVp2sAM3hxY-86WLswwcHDchhs70KYe9FOG7Dyugtu-88I1MWxzk9xdS3PeVbFmgsaW2M1U9ZSZZBXmmNvblpIK5jXjHljjUSIOye-38wJpShV1BOJjfEUSUIisP_NLsTmMbZ3Pb6bmdosVPIBPa03ZA</recordid><startdate>20160209</startdate><enddate>20160209</enddate><creator>Chwialkowski, Kacper</creator><creator>Strathmann, Heiko</creator><creator>Gretton, Arthur</creator><scope>EPD</scope><scope>GOX</scope></search><sort><creationdate>20160209</creationdate><title>A Kernel Test of Goodness of Fit</title><author>Chwialkowski, Kacper ; Strathmann, Heiko ; Gretton, Arthur</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a674-dcda59dd49c0f9a62fcba78d75fa55fcdc8006ee71605e7994494f382ccf40833</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2016</creationdate><topic>Statistics - Machine Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Chwialkowski, Kacper</creatorcontrib><creatorcontrib>Strathmann, Heiko</creatorcontrib><creatorcontrib>Gretton, Arthur</creatorcontrib><collection>arXiv Statistics</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Chwialkowski, Kacper</au><au>Strathmann, Heiko</au><au>Gretton, Arthur</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>A Kernel Test of Goodness of Fit</atitle><date>2016-02-09</date><risdate>2016</risdate><abstract>We propose a nonparametric statistical test for goodness-of-fit: given a set of samples, the test determines how likely it is that these were generated from a target density function. The measure of goodness-of-fit is a divergence constructed via Stein's method using functions from a Reproducing Kernel Hilbert Space. Our test statistic is based on an empirical estimate of this divergence, taking the form of a V-statistic in terms of the log gradients of the target density and the kernel. We derive a statistical test, both for i.i.d. and non-i.i.d. samples, where we estimate the null distribution quantiles using a wild bootstrap procedure. We apply our test to quantifying convergence of approximate Markov Chain Monte Carlo methods, statistical model criticism, and evaluating quality of fit vs model complexity in nonparametric density estimation.</abstract><doi>10.48550/arxiv.1602.02964</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.1602.02964
ispartof
issn
language eng
recordid cdi_arxiv_primary_1602_02964
source arXiv.org
subjects Statistics - Machine Learning
title A Kernel Test of Goodness of Fit
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-01T06%3A07%3A46IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=A%20Kernel%20Test%20of%20Goodness%20of%20Fit&rft.au=Chwialkowski,%20Kacper&rft.date=2016-02-09&rft_id=info:doi/10.48550/arxiv.1602.02964&rft_dat=%3Carxiv_GOX%3E1602_02964%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true