Local MALA-within-Gibbs for Bayesian image deblurring with total variation prior

We consider Bayesian inference for image deblurring with total variation (TV) prior. Since the posterior is analytically intractable, we resort to Markov chain Monte Carlo (MCMC) methods. However, since most MCMC methods significantly deteriorate in high dimensions, they are not suitable to handle h...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Flock, Rafael, Liu, Shuigen, Dong, Yiqiu, Tong, Xin T
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Flock, Rafael
Liu, Shuigen
Dong, Yiqiu
Tong, Xin T
description We consider Bayesian inference for image deblurring with total variation (TV) prior. Since the posterior is analytically intractable, we resort to Markov chain Monte Carlo (MCMC) methods. However, since most MCMC methods significantly deteriorate in high dimensions, they are not suitable to handle high resolution imaging problems. In this paper, we show how low-dimensional sampling can still be facilitated by exploiting the sparse conditional structure of the posterior. To this end, we make use of the local structures of the blurring operator and the TV prior by partitioning the image into rectangular blocks and employing a blocked Gibbs sampler with proposals stemming from the Metropolis-Hastings adjusted Langevin Algorithm (MALA). We prove that this MALA-within-Gibbs (MLwG) sampling algorithm has dimension-independent block acceptance rates and dimension-independent convergence rate. In order to apply the MALA proposals, we approximate the TV by a smoothed version, and show that the introduced approximation error is evenly distributed and dimension-independent. Since the posterior is a Gibbs density, we can use the Hammersley-Clifford Theorem to identify the posterior conditionals which only depend locally on the neighboring blocks. We outline computational strategies to evaluate the conditionals, which are the target densities in the Gibbs updates, locally and in parallel. In two numerical experiments, we validate the dimension-independent properties of the MLwG algorithm and demonstrate its superior performance over MALA.
doi_str_mv 10.48550/arxiv.2409.09810
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2409_09810</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2409_09810</sourcerecordid><originalsourceid>FETCH-arxiv_primary_2409_098103</originalsourceid><addsrcrecordid>eNpjYJA0NNAzsTA1NdBPLKrILNMzMjGw1DOwtDA04GQI8MlPTsxR8HX0cdQtzyzJyMzTdc9MSipWSMsvUnBKrEwtzkzMU8jMTUxPVUhJTcopLSrKzEtXAClVKMkvAWotSyzKTCzJzM9TKCjKzC_iYWBNS8wpTuWF0twM8m6uIc4eumC744FqchOLKuNBbogHu8GYsAoAuEk8Xg</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Local MALA-within-Gibbs for Bayesian image deblurring with total variation prior</title><source>arXiv.org</source><creator>Flock, Rafael ; Liu, Shuigen ; Dong, Yiqiu ; Tong, Xin T</creator><creatorcontrib>Flock, Rafael ; Liu, Shuigen ; Dong, Yiqiu ; Tong, Xin T</creatorcontrib><description>We consider Bayesian inference for image deblurring with total variation (TV) prior. Since the posterior is analytically intractable, we resort to Markov chain Monte Carlo (MCMC) methods. However, since most MCMC methods significantly deteriorate in high dimensions, they are not suitable to handle high resolution imaging problems. In this paper, we show how low-dimensional sampling can still be facilitated by exploiting the sparse conditional structure of the posterior. To this end, we make use of the local structures of the blurring operator and the TV prior by partitioning the image into rectangular blocks and employing a blocked Gibbs sampler with proposals stemming from the Metropolis-Hastings adjusted Langevin Algorithm (MALA). We prove that this MALA-within-Gibbs (MLwG) sampling algorithm has dimension-independent block acceptance rates and dimension-independent convergence rate. In order to apply the MALA proposals, we approximate the TV by a smoothed version, and show that the introduced approximation error is evenly distributed and dimension-independent. Since the posterior is a Gibbs density, we can use the Hammersley-Clifford Theorem to identify the posterior conditionals which only depend locally on the neighboring blocks. We outline computational strategies to evaluate the conditionals, which are the target densities in the Gibbs updates, locally and in parallel. In two numerical experiments, we validate the dimension-independent properties of the MLwG algorithm and demonstrate its superior performance over MALA.</description><identifier>DOI: 10.48550/arxiv.2409.09810</identifier><language>eng</language><subject>Computer Science - Numerical Analysis ; Mathematics - Numerical Analysis</subject><creationdate>2024-09</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2409.09810$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2409.09810$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Flock, Rafael</creatorcontrib><creatorcontrib>Liu, Shuigen</creatorcontrib><creatorcontrib>Dong, Yiqiu</creatorcontrib><creatorcontrib>Tong, Xin T</creatorcontrib><title>Local MALA-within-Gibbs for Bayesian image deblurring with total variation prior</title><description>We consider Bayesian inference for image deblurring with total variation (TV) prior. Since the posterior is analytically intractable, we resort to Markov chain Monte Carlo (MCMC) methods. However, since most MCMC methods significantly deteriorate in high dimensions, they are not suitable to handle high resolution imaging problems. In this paper, we show how low-dimensional sampling can still be facilitated by exploiting the sparse conditional structure of the posterior. To this end, we make use of the local structures of the blurring operator and the TV prior by partitioning the image into rectangular blocks and employing a blocked Gibbs sampler with proposals stemming from the Metropolis-Hastings adjusted Langevin Algorithm (MALA). We prove that this MALA-within-Gibbs (MLwG) sampling algorithm has dimension-independent block acceptance rates and dimension-independent convergence rate. In order to apply the MALA proposals, we approximate the TV by a smoothed version, and show that the introduced approximation error is evenly distributed and dimension-independent. Since the posterior is a Gibbs density, we can use the Hammersley-Clifford Theorem to identify the posterior conditionals which only depend locally on the neighboring blocks. We outline computational strategies to evaluate the conditionals, which are the target densities in the Gibbs updates, locally and in parallel. In two numerical experiments, we validate the dimension-independent properties of the MLwG algorithm and demonstrate its superior performance over MALA.</description><subject>Computer Science - Numerical Analysis</subject><subject>Mathematics - Numerical Analysis</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNpjYJA0NNAzsTA1NdBPLKrILNMzMjGw1DOwtDA04GQI8MlPTsxR8HX0cdQtzyzJyMzTdc9MSipWSMsvUnBKrEwtzkzMU8jMTUxPVUhJTcopLSrKzEtXAClVKMkvAWotSyzKTCzJzM9TKCjKzC_iYWBNS8wpTuWF0twM8m6uIc4eumC744FqchOLKuNBbogHu8GYsAoAuEk8Xg</recordid><startdate>20240915</startdate><enddate>20240915</enddate><creator>Flock, Rafael</creator><creator>Liu, Shuigen</creator><creator>Dong, Yiqiu</creator><creator>Tong, Xin T</creator><scope>AKY</scope><scope>AKZ</scope><scope>GOX</scope></search><sort><creationdate>20240915</creationdate><title>Local MALA-within-Gibbs for Bayesian image deblurring with total variation prior</title><author>Flock, Rafael ; Liu, Shuigen ; Dong, Yiqiu ; Tong, Xin T</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-arxiv_primary_2409_098103</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Numerical Analysis</topic><topic>Mathematics - Numerical Analysis</topic><toplevel>online_resources</toplevel><creatorcontrib>Flock, Rafael</creatorcontrib><creatorcontrib>Liu, Shuigen</creatorcontrib><creatorcontrib>Dong, Yiqiu</creatorcontrib><creatorcontrib>Tong, Xin T</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv Mathematics</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Flock, Rafael</au><au>Liu, Shuigen</au><au>Dong, Yiqiu</au><au>Tong, Xin T</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Local MALA-within-Gibbs for Bayesian image deblurring with total variation prior</atitle><date>2024-09-15</date><risdate>2024</risdate><abstract>We consider Bayesian inference for image deblurring with total variation (TV) prior. Since the posterior is analytically intractable, we resort to Markov chain Monte Carlo (MCMC) methods. However, since most MCMC methods significantly deteriorate in high dimensions, they are not suitable to handle high resolution imaging problems. In this paper, we show how low-dimensional sampling can still be facilitated by exploiting the sparse conditional structure of the posterior. To this end, we make use of the local structures of the blurring operator and the TV prior by partitioning the image into rectangular blocks and employing a blocked Gibbs sampler with proposals stemming from the Metropolis-Hastings adjusted Langevin Algorithm (MALA). We prove that this MALA-within-Gibbs (MLwG) sampling algorithm has dimension-independent block acceptance rates and dimension-independent convergence rate. In order to apply the MALA proposals, we approximate the TV by a smoothed version, and show that the introduced approximation error is evenly distributed and dimension-independent. Since the posterior is a Gibbs density, we can use the Hammersley-Clifford Theorem to identify the posterior conditionals which only depend locally on the neighboring blocks. We outline computational strategies to evaluate the conditionals, which are the target densities in the Gibbs updates, locally and in parallel. In two numerical experiments, we validate the dimension-independent properties of the MLwG algorithm and demonstrate its superior performance over MALA.</abstract><doi>10.48550/arxiv.2409.09810</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2409.09810
ispartof
issn
language eng
recordid cdi_arxiv_primary_2409_09810
source arXiv.org
subjects Computer Science - Numerical Analysis
Mathematics - Numerical Analysis
title Local MALA-within-Gibbs for Bayesian image deblurring with total variation prior
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-29T10%3A09%3A01IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Local%20MALA-within-Gibbs%20for%20Bayesian%20image%20deblurring%20with%20total%20variation%20prior&rft.au=Flock,%20Rafael&rft.date=2024-09-15&rft_id=info:doi/10.48550/arxiv.2409.09810&rft_dat=%3Carxiv_GOX%3E2409_09810%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true