A Kernel Test of Goodness of Fit
We propose a nonparametric statistical test for goodness-of-fit: given a set of samples, the test determines how likely it is that these were generated from a target density function. The measure of goodness-of-fit is a divergence constructed via Stein's method using functions from a Reproducin...
Gespeichert in:
Hauptverfasser: | , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Chwialkowski, Kacper Strathmann, Heiko Gretton, Arthur |
description | We propose a nonparametric statistical test for goodness-of-fit: given a set
of samples, the test determines how likely it is that these were generated from
a target density function. The measure of goodness-of-fit is a divergence
constructed via Stein's method using functions from a Reproducing Kernel
Hilbert Space. Our test statistic is based on an empirical estimate of this
divergence, taking the form of a V-statistic in terms of the log gradients of
the target density and the kernel. We derive a statistical test, both for
i.i.d. and non-i.i.d. samples, where we estimate the null distribution
quantiles using a wild bootstrap procedure. We apply our test to quantifying
convergence of approximate Markov Chain Monte Carlo methods, statistical model
criticism, and evaluating quality of fit vs model complexity in nonparametric
density estimation. |
doi_str_mv | 10.48550/arxiv.1602.02964 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_1602_02964</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>1602_02964</sourcerecordid><originalsourceid>FETCH-LOGICAL-a674-dcda59dd49c0f9a62fcba78d75fa55fcdc8006ee71605e7994494f382ccf40833</originalsourceid><addsrcrecordid>eNotzr0OgjAABOAuDgZ9ACf7AmDpf0dCBI0kLuyk9ichQTCUGH17BZ3uprsPgF2KEioZQwc9vtpnknKEE4QVp2sAM3hxY-86WLswwcHDchhs70KYe9FOG7Dyugtu-88I1MWxzk9xdS3PeVbFmgsaW2M1U9ZSZZBXmmNvblpIK5jXjHljjUSIOye-38wJpShV1BOJjfEUSUIisP_NLsTmMbZ3Pb6bmdosVPIBPa03ZA</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>A Kernel Test of Goodness of Fit</title><source>arXiv.org</source><creator>Chwialkowski, Kacper ; Strathmann, Heiko ; Gretton, Arthur</creator><creatorcontrib>Chwialkowski, Kacper ; Strathmann, Heiko ; Gretton, Arthur</creatorcontrib><description>We propose a nonparametric statistical test for goodness-of-fit: given a set
of samples, the test determines how likely it is that these were generated from
a target density function. The measure of goodness-of-fit is a divergence
constructed via Stein's method using functions from a Reproducing Kernel
Hilbert Space. Our test statistic is based on an empirical estimate of this
divergence, taking the form of a V-statistic in terms of the log gradients of
the target density and the kernel. We derive a statistical test, both for
i.i.d. and non-i.i.d. samples, where we estimate the null distribution
quantiles using a wild bootstrap procedure. We apply our test to quantifying
convergence of approximate Markov Chain Monte Carlo methods, statistical model
criticism, and evaluating quality of fit vs model complexity in nonparametric
density estimation.</description><identifier>DOI: 10.48550/arxiv.1602.02964</identifier><language>eng</language><subject>Statistics - Machine Learning</subject><creationdate>2016-02</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,782,887</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/1602.02964$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.1602.02964$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Chwialkowski, Kacper</creatorcontrib><creatorcontrib>Strathmann, Heiko</creatorcontrib><creatorcontrib>Gretton, Arthur</creatorcontrib><title>A Kernel Test of Goodness of Fit</title><description>We propose a nonparametric statistical test for goodness-of-fit: given a set
of samples, the test determines how likely it is that these were generated from
a target density function. The measure of goodness-of-fit is a divergence
constructed via Stein's method using functions from a Reproducing Kernel
Hilbert Space. Our test statistic is based on an empirical estimate of this
divergence, taking the form of a V-statistic in terms of the log gradients of
the target density and the kernel. We derive a statistical test, both for
i.i.d. and non-i.i.d. samples, where we estimate the null distribution
quantiles using a wild bootstrap procedure. We apply our test to quantifying
convergence of approximate Markov Chain Monte Carlo methods, statistical model
criticism, and evaluating quality of fit vs model complexity in nonparametric
density estimation.</description><subject>Statistics - Machine Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2016</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotzr0OgjAABOAuDgZ9ACf7AmDpf0dCBI0kLuyk9ichQTCUGH17BZ3uprsPgF2KEioZQwc9vtpnknKEE4QVp2sAM3hxY-86WLswwcHDchhs70KYe9FOG7Dyugtu-88I1MWxzk9xdS3PeVbFmgsaW2M1U9ZSZZBXmmNvblpIK5jXjHljjUSIOye-38wJpShV1BOJjfEUSUIisP_NLsTmMbZ3Pb6bmdosVPIBPa03ZA</recordid><startdate>20160209</startdate><enddate>20160209</enddate><creator>Chwialkowski, Kacper</creator><creator>Strathmann, Heiko</creator><creator>Gretton, Arthur</creator><scope>EPD</scope><scope>GOX</scope></search><sort><creationdate>20160209</creationdate><title>A Kernel Test of Goodness of Fit</title><author>Chwialkowski, Kacper ; Strathmann, Heiko ; Gretton, Arthur</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a674-dcda59dd49c0f9a62fcba78d75fa55fcdc8006ee71605e7994494f382ccf40833</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2016</creationdate><topic>Statistics - Machine Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Chwialkowski, Kacper</creatorcontrib><creatorcontrib>Strathmann, Heiko</creatorcontrib><creatorcontrib>Gretton, Arthur</creatorcontrib><collection>arXiv Statistics</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Chwialkowski, Kacper</au><au>Strathmann, Heiko</au><au>Gretton, Arthur</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>A Kernel Test of Goodness of Fit</atitle><date>2016-02-09</date><risdate>2016</risdate><abstract>We propose a nonparametric statistical test for goodness-of-fit: given a set
of samples, the test determines how likely it is that these were generated from
a target density function. The measure of goodness-of-fit is a divergence
constructed via Stein's method using functions from a Reproducing Kernel
Hilbert Space. Our test statistic is based on an empirical estimate of this
divergence, taking the form of a V-statistic in terms of the log gradients of
the target density and the kernel. We derive a statistical test, both for
i.i.d. and non-i.i.d. samples, where we estimate the null distribution
quantiles using a wild bootstrap procedure. We apply our test to quantifying
convergence of approximate Markov Chain Monte Carlo methods, statistical model
criticism, and evaluating quality of fit vs model complexity in nonparametric
density estimation.</abstract><doi>10.48550/arxiv.1602.02964</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.1602.02964 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_1602_02964 |
source | arXiv.org |
subjects | Statistics - Machine Learning |
title | A Kernel Test of Goodness of Fit |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-01T06%3A07%3A46IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=A%20Kernel%20Test%20of%20Goodness%20of%20Fit&rft.au=Chwialkowski,%20Kacper&rft.date=2016-02-09&rft_id=info:doi/10.48550/arxiv.1602.02964&rft_dat=%3Carxiv_GOX%3E1602_02964%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |