Visible and Infrared Image Fusion Using Encoder-Decoder Network

The aim of multispectral image fusion is to combine object or scene features of images with different spectral characteristics to increase the perceptual quality. In this paper, we present a novel learning-based solution to image fusion problem focusing on infrared and visible spectrum images. The p...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:arXiv.org 2024-12
Hauptverfasser: Ferhat Can Ataman, Akar, Gözde Bozdaği
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Ferhat Can Ataman
Akar, Gözde Bozdaği
description The aim of multispectral image fusion is to combine object or scene features of images with different spectral characteristics to increase the perceptual quality. In this paper, we present a novel learning-based solution to image fusion problem focusing on infrared and visible spectrum images. The proposed solution utilizes only convolution and pooling layers together with a loss function using no-reference quality metrics. The analysis is performed qualitatively and quantitatively on various datasets. The results show better performance than state-of-the-art methods. Also, the size of our network enables real-time performance on embedded devices. Project codes can be found at \url{https://github.com/ferhatcan/pyFusionSR}.
doi_str_mv 10.48550/arxiv.2412.08073
format Article
fullrecord <record><control><sourceid>proquest_arxiv</sourceid><recordid>TN_cdi_arxiv_primary_2412_08073</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3143450832</sourcerecordid><originalsourceid>FETCH-LOGICAL-a522-75ec825b667858c7070c9b60a74faf2931d98462e2461f163a1d061bc3fd71163</originalsourceid><addsrcrecordid>eNotj01Lw0AYhBdBsNT-AE8GPKe---5nTiK11ULRS_UaNslu2dpu6m7jx783pp5mBoZhHkKuKEy5FgJuTfz2n1PkFKegQbEzMkLGaK454gWZpLQFAJQKhWAjcvfmk692NjOhyZbBRRNtb_ZmY7NFl3wbstfkwyabh7ptbMwf7KDZsz1-tfH9kpw7s0t28q9jsl7M17OnfPXyuJzdr3IjEHMlbK1RVFIqLXStQEFdVBKM4s44LBhtCs0lWuSSOiqZoQ1IWtXMNYr2eUyuT7MDXXmIfm_iT_lHWQ6UfePm1DjE9qOz6Vhu2y6G_lPJKGdcgGbIfgGKglKf</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3143450832</pqid></control><display><type>article</type><title>Visible and Infrared Image Fusion Using Encoder-Decoder Network</title><source>arXiv.org</source><source>Free E- Journals</source><creator>Ferhat Can Ataman ; Akar, Gözde Bozdaği</creator><creatorcontrib>Ferhat Can Ataman ; Akar, Gözde Bozdaği</creatorcontrib><description>The aim of multispectral image fusion is to combine object or scene features of images with different spectral characteristics to increase the perceptual quality. In this paper, we present a novel learning-based solution to image fusion problem focusing on infrared and visible spectrum images. The proposed solution utilizes only convolution and pooling layers together with a loss function using no-reference quality metrics. The analysis is performed qualitatively and quantitatively on various datasets. The results show better performance than state-of-the-art methods. Also, the size of our network enables real-time performance on embedded devices. Project codes can be found at \url{https://github.com/ferhatcan/pyFusionSR}.</description><identifier>EISSN: 2331-8422</identifier><identifier>DOI: 10.48550/arxiv.2412.08073</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Computer Science - Computer Vision and Pattern Recognition ; Computer Science - Learning ; Computer vision ; Encoders-Decoders ; Image quality ; Infrared analysis ; Infrared imagery ; Real time ; Visible spectrum</subject><ispartof>arXiv.org, 2024-12</ispartof><rights>2024. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,780,881,27902</link.rule.ids><backlink>$$Uhttps://doi.org/10.48550/arXiv.2412.08073$$DView paper in arXiv$$Hfree_for_read</backlink><backlink>$$Uhttps://doi.org/10.1109/ICIP42928.2021.9506740$$DView published paper (Access to full text may be restricted)$$Hfree_for_read</backlink></links><search><creatorcontrib>Ferhat Can Ataman</creatorcontrib><creatorcontrib>Akar, Gözde Bozdaği</creatorcontrib><title>Visible and Infrared Image Fusion Using Encoder-Decoder Network</title><title>arXiv.org</title><description>The aim of multispectral image fusion is to combine object or scene features of images with different spectral characteristics to increase the perceptual quality. In this paper, we present a novel learning-based solution to image fusion problem focusing on infrared and visible spectrum images. The proposed solution utilizes only convolution and pooling layers together with a loss function using no-reference quality metrics. The analysis is performed qualitatively and quantitatively on various datasets. The results show better performance than state-of-the-art methods. Also, the size of our network enables real-time performance on embedded devices. Project codes can be found at \url{https://github.com/ferhatcan/pyFusionSR}.</description><subject>Computer Science - Computer Vision and Pattern Recognition</subject><subject>Computer Science - Learning</subject><subject>Computer vision</subject><subject>Encoders-Decoders</subject><subject>Image quality</subject><subject>Infrared analysis</subject><subject>Infrared imagery</subject><subject>Real time</subject><subject>Visible spectrum</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>BENPR</sourceid><sourceid>GOX</sourceid><recordid>eNotj01Lw0AYhBdBsNT-AE8GPKe---5nTiK11ULRS_UaNslu2dpu6m7jx783pp5mBoZhHkKuKEy5FgJuTfz2n1PkFKegQbEzMkLGaK454gWZpLQFAJQKhWAjcvfmk692NjOhyZbBRRNtb_ZmY7NFl3wbstfkwyabh7ptbMwf7KDZsz1-tfH9kpw7s0t28q9jsl7M17OnfPXyuJzdr3IjEHMlbK1RVFIqLXStQEFdVBKM4s44LBhtCs0lWuSSOiqZoQ1IWtXMNYr2eUyuT7MDXXmIfm_iT_lHWQ6UfePm1DjE9qOz6Vhu2y6G_lPJKGdcgGbIfgGKglKf</recordid><startdate>20241211</startdate><enddate>20241211</enddate><creator>Ferhat Can Ataman</creator><creator>Akar, Gözde Bozdaği</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20241211</creationdate><title>Visible and Infrared Image Fusion Using Encoder-Decoder Network</title><author>Ferhat Can Ataman ; Akar, Gözde Bozdaği</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a522-75ec825b667858c7070c9b60a74faf2931d98462e2461f163a1d061bc3fd71163</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Computer Vision and Pattern Recognition</topic><topic>Computer Science - Learning</topic><topic>Computer vision</topic><topic>Encoders-Decoders</topic><topic>Image quality</topic><topic>Infrared analysis</topic><topic>Infrared imagery</topic><topic>Real time</topic><topic>Visible spectrum</topic><toplevel>online_resources</toplevel><creatorcontrib>Ferhat Can Ataman</creatorcontrib><creatorcontrib>Akar, Gözde Bozdaği</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection><collection>arXiv Computer Science</collection><collection>arXiv.org</collection><jtitle>arXiv.org</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Ferhat Can Ataman</au><au>Akar, Gözde Bozdaği</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Visible and Infrared Image Fusion Using Encoder-Decoder Network</atitle><jtitle>arXiv.org</jtitle><date>2024-12-11</date><risdate>2024</risdate><eissn>2331-8422</eissn><abstract>The aim of multispectral image fusion is to combine object or scene features of images with different spectral characteristics to increase the perceptual quality. In this paper, we present a novel learning-based solution to image fusion problem focusing on infrared and visible spectrum images. The proposed solution utilizes only convolution and pooling layers together with a loss function using no-reference quality metrics. The analysis is performed qualitatively and quantitatively on various datasets. The results show better performance than state-of-the-art methods. Also, the size of our network enables real-time performance on embedded devices. Project codes can be found at \url{https://github.com/ferhatcan/pyFusionSR}.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><doi>10.48550/arxiv.2412.08073</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2024-12
issn 2331-8422
language eng
recordid cdi_arxiv_primary_2412_08073
source arXiv.org; Free E- Journals
subjects Computer Science - Computer Vision and Pattern Recognition
Computer Science - Learning
Computer vision
Encoders-Decoders
Image quality
Infrared analysis
Infrared imagery
Real time
Visible spectrum
title Visible and Infrared Image Fusion Using Encoder-Decoder Network
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-12T17%3A34%3A48IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_arxiv&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Visible%20and%20Infrared%20Image%20Fusion%20Using%20Encoder-Decoder%20Network&rft.jtitle=arXiv.org&rft.au=Ferhat%20Can%20Ataman&rft.date=2024-12-11&rft.eissn=2331-8422&rft_id=info:doi/10.48550/arxiv.2412.08073&rft_dat=%3Cproquest_arxiv%3E3143450832%3C/proquest_arxiv%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3143450832&rft_id=info:pmid/&rfr_iscdi=true