SAFNet: Selective Alignment Fusion Network for Efficient HDR Imaging

Multi-exposure High Dynamic Range (HDR) imaging is a challenging task when facing truncated texture and complex motion. Existing deep learning-based methods have achieved great success by either following the alignment and fusion pipeline or utilizing attention mechanism. However, the large computat...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:arXiv.org 2024-07
Hauptverfasser: Kong, Lingtong, Li, Bo, Xiong, Yike, Zhang, Hao, Gu, Hong, Chen, Jinwei
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Kong, Lingtong
Li, Bo
Xiong, Yike
Zhang, Hao
Gu, Hong
Chen, Jinwei
description Multi-exposure High Dynamic Range (HDR) imaging is a challenging task when facing truncated texture and complex motion. Existing deep learning-based methods have achieved great success by either following the alignment and fusion pipeline or utilizing attention mechanism. However, the large computation cost and inference delay hinder them from deploying on resource limited devices. In this paper, to achieve better efficiency, a novel Selective Alignment Fusion Network (SAFNet) for HDR imaging is proposed. After extracting pyramid features, it jointly refines valuable area masks and cross-exposure motion in selected regions with shared decoders, and then fuses high quality HDR image in an explicit way. This approach can focus the model on finding valuable regions while estimating their easily detectable and meaningful motion. For further detail enhancement, a lightweight refine module is introduced which enjoys privileges from previous optical flow, selection masks and initial prediction. Moreover, to facilitate learning on samples with large motion, a new window partition cropping method is presented during training. Experiments on public and newly developed challenging datasets show that proposed SAFNet not only exceeds previous SOTA competitors quantitatively and qualitatively, but also runs order of magnitude faster. Code and dataset is available at https://github.com/ltkong218/SAFNet.
format Article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_3084094470</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3084094470</sourcerecordid><originalsourceid>FETCH-proquest_journals_30840944703</originalsourceid><addsrcrecordid>eNqNysEKgjAcgPERBEn5Dn_oLKxtpnWTVOzSIbuLyDZmutU26_Uz6AE6fYfft0ABoXQXpYyQFQqd6zHGZJ-QOKYByuusvHB_hJoPvPPqxSEblNQj1x7KySmjYfa3sXcQxkIhhOrUF6v8CuexlUrLDVqKdnA8_HWNtmVxO1XRw5rnxJ1vejNZPVNDccrwgbEE0_-uD6g-OWE</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3084094470</pqid></control><display><type>article</type><title>SAFNet: Selective Alignment Fusion Network for Efficient HDR Imaging</title><source>Free E- Journals</source><creator>Kong, Lingtong ; Li, Bo ; Xiong, Yike ; Zhang, Hao ; Gu, Hong ; Chen, Jinwei</creator><creatorcontrib>Kong, Lingtong ; Li, Bo ; Xiong, Yike ; Zhang, Hao ; Gu, Hong ; Chen, Jinwei</creatorcontrib><description>Multi-exposure High Dynamic Range (HDR) imaging is a challenging task when facing truncated texture and complex motion. Existing deep learning-based methods have achieved great success by either following the alignment and fusion pipeline or utilizing attention mechanism. However, the large computation cost and inference delay hinder them from deploying on resource limited devices. In this paper, to achieve better efficiency, a novel Selective Alignment Fusion Network (SAFNet) for HDR imaging is proposed. After extracting pyramid features, it jointly refines valuable area masks and cross-exposure motion in selected regions with shared decoders, and then fuses high quality HDR image in an explicit way. This approach can focus the model on finding valuable regions while estimating their easily detectable and meaningful motion. For further detail enhancement, a lightweight refine module is introduced which enjoys privileges from previous optical flow, selection masks and initial prediction. Moreover, to facilitate learning on samples with large motion, a new window partition cropping method is presented during training. Experiments on public and newly developed challenging datasets show that proposed SAFNet not only exceeds previous SOTA competitors quantitatively and qualitatively, but also runs order of magnitude faster. Code and dataset is available at https://github.com/ltkong218/SAFNet.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Alignment ; Datasets ; Decoders ; Deep learning ; Image enhancement ; Image quality ; Masks ; Optical flow (image analysis)</subject><ispartof>arXiv.org, 2024-07</ispartof><rights>2024. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>780,784</link.rule.ids></links><search><creatorcontrib>Kong, Lingtong</creatorcontrib><creatorcontrib>Li, Bo</creatorcontrib><creatorcontrib>Xiong, Yike</creatorcontrib><creatorcontrib>Zhang, Hao</creatorcontrib><creatorcontrib>Gu, Hong</creatorcontrib><creatorcontrib>Chen, Jinwei</creatorcontrib><title>SAFNet: Selective Alignment Fusion Network for Efficient HDR Imaging</title><title>arXiv.org</title><description>Multi-exposure High Dynamic Range (HDR) imaging is a challenging task when facing truncated texture and complex motion. Existing deep learning-based methods have achieved great success by either following the alignment and fusion pipeline or utilizing attention mechanism. However, the large computation cost and inference delay hinder them from deploying on resource limited devices. In this paper, to achieve better efficiency, a novel Selective Alignment Fusion Network (SAFNet) for HDR imaging is proposed. After extracting pyramid features, it jointly refines valuable area masks and cross-exposure motion in selected regions with shared decoders, and then fuses high quality HDR image in an explicit way. This approach can focus the model on finding valuable regions while estimating their easily detectable and meaningful motion. For further detail enhancement, a lightweight refine module is introduced which enjoys privileges from previous optical flow, selection masks and initial prediction. Moreover, to facilitate learning on samples with large motion, a new window partition cropping method is presented during training. Experiments on public and newly developed challenging datasets show that proposed SAFNet not only exceeds previous SOTA competitors quantitatively and qualitatively, but also runs order of magnitude faster. Code and dataset is available at https://github.com/ltkong218/SAFNet.</description><subject>Alignment</subject><subject>Datasets</subject><subject>Decoders</subject><subject>Deep learning</subject><subject>Image enhancement</subject><subject>Image quality</subject><subject>Masks</subject><subject>Optical flow (image analysis)</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><recordid>eNqNysEKgjAcgPERBEn5Dn_oLKxtpnWTVOzSIbuLyDZmutU26_Uz6AE6fYfft0ABoXQXpYyQFQqd6zHGZJ-QOKYByuusvHB_hJoPvPPqxSEblNQj1x7KySmjYfa3sXcQxkIhhOrUF6v8CuexlUrLDVqKdnA8_HWNtmVxO1XRw5rnxJ1vejNZPVNDccrwgbEE0_-uD6g-OWE</recordid><startdate>20240723</startdate><enddate>20240723</enddate><creator>Kong, Lingtong</creator><creator>Li, Bo</creator><creator>Xiong, Yike</creator><creator>Zhang, Hao</creator><creator>Gu, Hong</creator><creator>Chen, Jinwei</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20240723</creationdate><title>SAFNet: Selective Alignment Fusion Network for Efficient HDR Imaging</title><author>Kong, Lingtong ; Li, Bo ; Xiong, Yike ; Zhang, Hao ; Gu, Hong ; Chen, Jinwei</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_30840944703</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Alignment</topic><topic>Datasets</topic><topic>Decoders</topic><topic>Deep learning</topic><topic>Image enhancement</topic><topic>Image quality</topic><topic>Masks</topic><topic>Optical flow (image analysis)</topic><toplevel>online_resources</toplevel><creatorcontrib>Kong, Lingtong</creatorcontrib><creatorcontrib>Li, Bo</creatorcontrib><creatorcontrib>Xiong, Yike</creatorcontrib><creatorcontrib>Zhang, Hao</creatorcontrib><creatorcontrib>Gu, Hong</creatorcontrib><creatorcontrib>Chen, Jinwei</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Kong, Lingtong</au><au>Li, Bo</au><au>Xiong, Yike</au><au>Zhang, Hao</au><au>Gu, Hong</au><au>Chen, Jinwei</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>SAFNet: Selective Alignment Fusion Network for Efficient HDR Imaging</atitle><jtitle>arXiv.org</jtitle><date>2024-07-23</date><risdate>2024</risdate><eissn>2331-8422</eissn><abstract>Multi-exposure High Dynamic Range (HDR) imaging is a challenging task when facing truncated texture and complex motion. Existing deep learning-based methods have achieved great success by either following the alignment and fusion pipeline or utilizing attention mechanism. However, the large computation cost and inference delay hinder them from deploying on resource limited devices. In this paper, to achieve better efficiency, a novel Selective Alignment Fusion Network (SAFNet) for HDR imaging is proposed. After extracting pyramid features, it jointly refines valuable area masks and cross-exposure motion in selected regions with shared decoders, and then fuses high quality HDR image in an explicit way. This approach can focus the model on finding valuable regions while estimating their easily detectable and meaningful motion. For further detail enhancement, a lightweight refine module is introduced which enjoys privileges from previous optical flow, selection masks and initial prediction. Moreover, to facilitate learning on samples with large motion, a new window partition cropping method is presented during training. Experiments on public and newly developed challenging datasets show that proposed SAFNet not only exceeds previous SOTA competitors quantitatively and qualitatively, but also runs order of magnitude faster. Code and dataset is available at https://github.com/ltkong218/SAFNet.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2024-07
issn 2331-8422
language eng
recordid cdi_proquest_journals_3084094470
source Free E- Journals
subjects Alignment
Datasets
Decoders
Deep learning
Image enhancement
Image quality
Masks
Optical flow (image analysis)
title SAFNet: Selective Alignment Fusion Network for Efficient HDR Imaging
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-06T02%3A30%3A45IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=SAFNet:%20Selective%20Alignment%20Fusion%20Network%20for%20Efficient%20HDR%20Imaging&rft.jtitle=arXiv.org&rft.au=Kong,%20Lingtong&rft.date=2024-07-23&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E3084094470%3C/proquest%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3084094470&rft_id=info:pmid/&rfr_iscdi=true